2026-03-10T09:57:00.705 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-10T09:57:00.710 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T09:57:00.727 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/992 branch: squid description: orch/cephadm/smoke-roleless/{0-distro/centos_9.stream_runc 1-start 2-services/nfs-ingress-rgw-bucket 3-final} email: null first_in_suite: false flavor: default job_id: '992' last_in_suite: false machine_type: vps name: kyr-2026-03-10_01:00:38-orch-squid-none-default-vps no_nested_subset: false openstack: - volumes: count: 4 size: 10 os_type: centos os_version: 9.stream overrides: admin_socket: branch: squid ansible.cephlab: branch: main skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: timezone: UTC ceph: conf: mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 osd shutdown pgref assert: true flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - CEPHADM_DAEMON_PLACE_FAIL - CEPHADM_FAILED_DAEMON log-only-match: - CEPHADM_ sha1: e911bdebe5c8faa3800735d1568fcdca65db60df ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} install: ceph: flavor: default sha1: e911bdebe5c8faa3800735d1568fcdca65db60df extra_system_packages: deb: - python3-xmltodict - python3-jmespath rpm: - bzip2 - perl-Test-Harness - python3-xmltodict - python3-jmespath selinux: allowlist: - scontext=system_u:system_r:logrotate_t:s0 workunit: branch: tt-squid sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - client.0 - - host.b - client.1 seed: 8043 sha1: e911bdebe5c8faa3800735d1568fcdca65db60df sleep_before_teardown: 0 subset: 1/64 suite: orch suite_branch: tt-squid suite_path: /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b targets: vm06.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPA7e0nk4bsu/gBjGNrapm0rlX8aKz4mUMmMA+sFQRr5UDO+06xFmmdtUhlfcVqCbr/XdxcXPAa0UOEJNQHA+WU= vm09.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBFyKtjkdUF3uvMWWQDklMKEqk/UyTM0kl3XcoP9BaPGT+rTVRf3G4JAZ7kwSf/LJlgjbbcFoZjiDT8AuUA7agzc= tasks: - pexec: all: - sudo dnf remove nvme-cli -y - sudo dnf install runc nvmetcli nvme-cli -y - sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf - sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf - cephadm: roleless: true - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - vip: null - cephadm.shell: host.a: - ceph orch device ls --refresh - vip.exec: all-hosts: - systemctl stop nfs-server - cephadm.shell: host.a: - ceph orch apply rgw foorgw --port 8800 - ceph nfs cluster create foo --ingress --virtual-ip {{VIP0}}/{{VIPPREFIXLEN}} - vip.exec: host.a: - dnf install -y python3-boto3 || apt install -y python3-boto3 - /home/ubuntu/cephtest/cephadm shell radosgw-admin user create --uid foouser --display-name foo > /tmp/user.json - python: host.a: "import boto3\nimport json\n\nwith open('/tmp/user.json', 'rt') as f:\n\ \ info = json.loads(f.read())\ns3 = boto3.resource(\n 's3',\n aws_access_key_id=info['keys'][0]['access_key'],\n\ \ aws_secret_access_key=info['keys'][0]['secret_key'],\n endpoint_url='http://localhost:8800',\n\ )\nbucket = s3.Bucket('foobucket')\nbucket.create()\nbucket.put_object(Key='myobject',\ \ Body='thebody')\n" - cephadm.shell: host.a: - ceph nfs export create rgw --bucket foobucket --cluster-id foo --pseudo-path /foobucket - cephadm.wait_for_service: service: nfs.foo - cephadm.wait_for_service: service: ingress.nfs.foo - vip.exec: host.a: - mkdir /mnt/foo - sleep 5 - mount -t nfs {{VIP0}}:/foobucket /mnt/foo - find /mnt/foo -ls - grep thebody /mnt/foo/myobject - echo test > /mnt/foo/newobject - sync - python: host.a: "import boto3\nimport json\nfrom io import BytesIO\n\nwith open('/tmp/user.json',\ \ 'rt') as f:\n info = json.loads(f.read())\ns3 = boto3.resource(\n 's3',\n\ \ aws_access_key_id=info['keys'][0]['access_key'],\n aws_secret_access_key=info['keys'][0]['secret_key'],\n\ \ endpoint_url='http://localhost:8800',\n)\nbucket = s3.Bucket('foobucket')\n\ data = BytesIO()\nbucket.download_fileobj(Fileobj=data, Key='newobject')\nprint(data.getvalue())\n\ assert data.getvalue().decode() == 'test\\n'\n" - vip.exec: host.a: - umount /mnt/foo - cephadm.shell: host.a: - ceph nfs export rm foo /foobucket - ceph nfs cluster rm foo - cephadm.shell: host.a: - stat -c '%u %g' /var/log/ceph | grep '167 167' - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - ceph orch ls | grep '^osd.all-available-devices ' teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-10_01:00:38 tube: vps user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.611473 2026-03-10T09:57:00.727 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa; will attempt to use it 2026-03-10T09:57:00.728 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_75a68fd8ca3f918fe9466b4c0bb385b7fc260a9b/qa/tasks 2026-03-10T09:57:00.728 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-10T09:57:00.728 INFO:teuthology.task.internal:Checking packages... 2026-03-10T09:57:00.728 INFO:teuthology.task.internal:Checking packages for os_type 'centos', flavor 'default' and ceph hash 'e911bdebe5c8faa3800735d1568fcdca65db60df' 2026-03-10T09:57:00.728 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-10T09:57:00.728 INFO:teuthology.packaging:ref: None 2026-03-10T09:57:00.728 INFO:teuthology.packaging:tag: None 2026-03-10T09:57:00.728 INFO:teuthology.packaging:branch: squid 2026-03-10T09:57:00.728 INFO:teuthology.packaging:sha1: e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T09:57:00.728 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&ref=squid 2026-03-10T09:57:01.439 INFO:teuthology.task.internal:Found packages for ceph version 19.2.3-678.ge911bdeb 2026-03-10T09:57:01.440 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-10T09:57:01.441 INFO:teuthology.task.internal:no buildpackages task found 2026-03-10T09:57:01.441 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-10T09:57:01.441 INFO:teuthology.task.internal:Saving configuration 2026-03-10T09:57:01.446 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-10T09:57:01.447 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-10T09:57:01.453 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm06.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/992', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 09:55:46.999898', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:06', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBPA7e0nk4bsu/gBjGNrapm0rlX8aKz4mUMmMA+sFQRr5UDO+06xFmmdtUhlfcVqCbr/XdxcXPAa0UOEJNQHA+WU='} 2026-03-10T09:57:01.458 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm09.local', 'description': '/archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/992', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'centos', 'os_version': '9.stream', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-10 09:55:46.999480', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:09', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBFyKtjkdUF3uvMWWQDklMKEqk/UyTM0kl3XcoP9BaPGT+rTVRf3G4JAZ7kwSf/LJlgjbbcFoZjiDT8AuUA7agzc='} 2026-03-10T09:57:01.458 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-10T09:57:01.458 INFO:teuthology.task.internal:roles: ubuntu@vm06.local - ['host.a', 'client.0'] 2026-03-10T09:57:01.458 INFO:teuthology.task.internal:roles: ubuntu@vm09.local - ['host.b', 'client.1'] 2026-03-10T09:57:01.458 INFO:teuthology.run_tasks:Running task console_log... 2026-03-10T09:57:01.464 DEBUG:teuthology.task.console_log:vm06 does not support IPMI; excluding 2026-03-10T09:57:01.469 DEBUG:teuthology.task.console_log:vm09 does not support IPMI; excluding 2026-03-10T09:57:01.469 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7ff456276170>, signals=[15]) 2026-03-10T09:57:01.469 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-10T09:57:01.470 INFO:teuthology.task.internal:Opening connections... 2026-03-10T09:57:01.470 DEBUG:teuthology.task.internal:connecting to ubuntu@vm06.local 2026-03-10T09:57:01.470 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T09:57:01.533 DEBUG:teuthology.task.internal:connecting to ubuntu@vm09.local 2026-03-10T09:57:01.534 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T09:57:01.591 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-10T09:57:01.593 DEBUG:teuthology.orchestra.run.vm06:> uname -m 2026-03-10T09:57:01.637 INFO:teuthology.orchestra.run.vm06.stdout:x86_64 2026-03-10T09:57:01.637 DEBUG:teuthology.orchestra.run.vm06:> cat /etc/os-release 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:NAME="CentOS Stream" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:VERSION="9" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:ID="centos" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:ID_LIKE="rhel fedora" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:VERSION_ID="9" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:PLATFORM_ID="platform:el9" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:ANSI_COLOR="0;31" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:LOGO="fedora-logo-icon" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:HOME_URL="https://centos.org/" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T09:57:01.693 INFO:teuthology.orchestra.run.vm06.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T09:57:01.694 INFO:teuthology.lock.ops:Updating vm06.local on lock server 2026-03-10T09:57:01.698 DEBUG:teuthology.orchestra.run.vm09:> uname -m 2026-03-10T09:57:01.712 INFO:teuthology.orchestra.run.vm09.stdout:x86_64 2026-03-10T09:57:01.712 DEBUG:teuthology.orchestra.run.vm09:> cat /etc/os-release 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:NAME="CentOS Stream" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:VERSION="9" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:ID="centos" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:ID_LIKE="rhel fedora" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:VERSION_ID="9" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:PLATFORM_ID="platform:el9" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:PRETTY_NAME="CentOS Stream 9" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:ANSI_COLOR="0;31" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:LOGO="fedora-logo-icon" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:CPE_NAME="cpe:/o:centos:centos:9" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:HOME_URL="https://centos.org/" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:BUG_REPORT_URL="https://issues.redhat.com/" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:REDHAT_SUPPORT_PRODUCT="Red Hat Enterprise Linux 9" 2026-03-10T09:57:01.766 INFO:teuthology.orchestra.run.vm09.stdout:REDHAT_SUPPORT_PRODUCT_VERSION="CentOS Stream" 2026-03-10T09:57:01.766 INFO:teuthology.lock.ops:Updating vm09.local on lock server 2026-03-10T09:57:01.770 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-10T09:57:01.772 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-10T09:57:01.772 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-10T09:57:01.773 DEBUG:teuthology.orchestra.run.vm06:> test '!' -e /home/ubuntu/cephtest 2026-03-10T09:57:01.774 DEBUG:teuthology.orchestra.run.vm09:> test '!' -e /home/ubuntu/cephtest 2026-03-10T09:57:01.820 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-10T09:57:01.821 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-10T09:57:01.821 DEBUG:teuthology.orchestra.run.vm06:> test -z $(ls -A /var/lib/ceph) 2026-03-10T09:57:01.829 DEBUG:teuthology.orchestra.run.vm09:> test -z $(ls -A /var/lib/ceph) 2026-03-10T09:57:01.841 INFO:teuthology.orchestra.run.vm06.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T09:57:01.874 INFO:teuthology.orchestra.run.vm09.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-10T09:57:01.875 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-10T09:57:01.882 DEBUG:teuthology.orchestra.run.vm06:> test -e /ceph-qa-ready 2026-03-10T09:57:01.895 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T09:57:02.095 DEBUG:teuthology.orchestra.run.vm09:> test -e /ceph-qa-ready 2026-03-10T09:57:02.109 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T09:57:02.285 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-10T09:57:02.286 INFO:teuthology.task.internal:Creating test directory... 2026-03-10T09:57:02.286 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T09:57:02.288 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-10T09:57:02.304 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-10T09:57:02.306 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-10T09:57:02.307 INFO:teuthology.task.internal:Creating archive directory... 2026-03-10T09:57:02.307 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T09:57:02.345 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-10T09:57:02.363 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-10T09:57:02.364 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-10T09:57:02.364 DEBUG:teuthology.orchestra.run.vm06:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T09:57:02.413 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T09:57:02.413 DEBUG:teuthology.orchestra.run.vm09:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-10T09:57:02.427 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T09:57:02.427 DEBUG:teuthology.orchestra.run.vm06:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T09:57:02.455 DEBUG:teuthology.orchestra.run.vm09:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-10T09:57:02.476 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T09:57:02.486 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T09:57:02.491 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T09:57:02.500 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-10T09:57:02.501 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-10T09:57:02.503 INFO:teuthology.task.internal:Configuring sudo... 2026-03-10T09:57:02.503 DEBUG:teuthology.orchestra.run.vm06:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T09:57:02.530 DEBUG:teuthology.orchestra.run.vm09:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-10T09:57:02.565 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-10T09:57:02.567 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-10T09:57:02.568 DEBUG:teuthology.orchestra.run.vm06:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T09:57:02.594 DEBUG:teuthology.orchestra.run.vm09:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-10T09:57:02.620 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T09:57:02.668 DEBUG:teuthology.orchestra.run.vm06:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T09:57:02.725 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T09:57:02.725 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T09:57:02.783 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T09:57:02.804 DEBUG:teuthology.orchestra.run.vm09:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T09:57:02.859 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T09:57:02.859 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-10T09:57:02.916 DEBUG:teuthology.orchestra.run.vm06:> sudo service rsyslog restart 2026-03-10T09:57:02.918 DEBUG:teuthology.orchestra.run.vm09:> sudo service rsyslog restart 2026-03-10T09:57:02.943 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T09:57:02.984 INFO:teuthology.orchestra.run.vm09.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T09:57:03.304 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-10T09:57:03.305 INFO:teuthology.task.internal:Starting timer... 2026-03-10T09:57:03.306 INFO:teuthology.run_tasks:Running task pcp... 2026-03-10T09:57:03.308 INFO:teuthology.run_tasks:Running task selinux... 2026-03-10T09:57:03.310 DEBUG:teuthology.task:Applying overrides for task selinux: {'allowlist': ['scontext=system_u:system_r:logrotate_t:s0']} 2026-03-10T09:57:03.310 INFO:teuthology.task.selinux:Excluding vm06: VMs are not yet supported 2026-03-10T09:57:03.310 INFO:teuthology.task.selinux:Excluding vm09: VMs are not yet supported 2026-03-10T09:57:03.310 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-10T09:57:03.310 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-10T09:57:03.310 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-10T09:57:03.310 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-10T09:57:03.312 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'timezone': 'UTC'}} 2026-03-10T09:57:03.312 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/ceph/ceph-cm-ansible.git 2026-03-10T09:57:03.313 INFO:teuthology.repo_utils:Fetching github.com_ceph_ceph-cm-ansible_main from origin 2026-03-10T09:57:04.000 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main to origin/main 2026-03-10T09:57:04.006 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-10T09:57:04.006 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "timezone": "UTC"}' -i /tmp/teuth_ansible_inventoryw3q2t5hf --limit vm06.local,vm09.local /home/teuthos/src/github.com_ceph_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-10T09:59:40.369 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm06.local'), Remote(name='ubuntu@vm09.local')] 2026-03-10T09:59:40.369 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm06.local' 2026-03-10T09:59:40.369 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm06.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T09:59:40.433 DEBUG:teuthology.orchestra.run.vm06:> true 2026-03-10T09:59:40.512 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm06.local' 2026-03-10T09:59:40.512 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm09.local' 2026-03-10T09:59:40.512 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm09.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-10T09:59:40.572 DEBUG:teuthology.orchestra.run.vm09:> true 2026-03-10T09:59:40.650 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm09.local' 2026-03-10T09:59:40.651 INFO:teuthology.run_tasks:Running task clock... 2026-03-10T09:59:40.654 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-10T09:59:40.654 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T09:59:40.654 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T09:59:40.657 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-10T09:59:40.657 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T09:59:40.693 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T09:59:40.713 INFO:teuthology.orchestra.run.vm06.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T09:59:40.716 INFO:teuthology.orchestra.run.vm09.stderr:Failed to stop ntp.service: Unit ntp.service not loaded. 2026-03-10T09:59:40.729 INFO:teuthology.orchestra.run.vm09.stderr:Failed to stop ntpd.service: Unit ntpd.service not loaded. 2026-03-10T09:59:40.743 INFO:teuthology.orchestra.run.vm06.stderr:sudo: ntpd: command not found 2026-03-10T09:59:40.753 INFO:teuthology.orchestra.run.vm09.stderr:sudo: ntpd: command not found 2026-03-10T09:59:40.757 INFO:teuthology.orchestra.run.vm06.stdout:506 Cannot talk to daemon 2026-03-10T09:59:40.765 INFO:teuthology.orchestra.run.vm09.stdout:506 Cannot talk to daemon 2026-03-10T09:59:40.773 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T09:59:40.778 INFO:teuthology.orchestra.run.vm09.stderr:Failed to start ntp.service: Unit ntp.service not found. 2026-03-10T09:59:40.791 INFO:teuthology.orchestra.run.vm09.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T09:59:40.792 INFO:teuthology.orchestra.run.vm06.stderr:Failed to start ntpd.service: Unit ntpd.service not found. 2026-03-10T09:59:40.840 INFO:teuthology.orchestra.run.vm09.stderr:bash: line 1: ntpq: command not found 2026-03-10T09:59:40.842 INFO:teuthology.orchestra.run.vm09.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T09:59:40.842 INFO:teuthology.orchestra.run.vm09.stdout:=============================================================================== 2026-03-10T09:59:40.844 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-10T09:59:40.846 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T09:59:40.846 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-10T09:59:40.846 INFO:teuthology.run_tasks:Running task pexec... 2026-03-10T09:59:40.848 INFO:teuthology.task.pexec:Executing custom commands... 2026-03-10T09:59:40.849 DEBUG:teuthology.orchestra.run.vm06:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T09:59:40.849 DEBUG:teuthology.orchestra.run.vm09:> TESTDIR=/home/ubuntu/cephtest bash -s 2026-03-10T09:59:40.884 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo dnf remove nvme-cli -y 2026-03-10T09:59:40.884 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T09:59:40.884 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T09:59:40.884 DEBUG:teuthology.task.pexec:ubuntu@vm09.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T09:59:40.884 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm09.local 2026-03-10T09:59:40.884 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T09:59:40.884 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T09:59:40.884 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T09:59:40.884 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T09:59:40.887 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf remove nvme-cli -y 2026-03-10T09:59:40.888 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T09:59:40.888 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T09:59:40.888 DEBUG:teuthology.task.pexec:ubuntu@vm06.local< sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T09:59:40.888 INFO:teuthology.task.pexec:Running commands on host ubuntu@vm06.local 2026-03-10T09:59:40.888 INFO:teuthology.task.pexec:sudo dnf remove nvme-cli -y 2026-03-10T09:59:40.888 INFO:teuthology.task.pexec:sudo dnf install runc nvmetcli nvme-cli -y 2026-03-10T09:59:40.888 INFO:teuthology.task.pexec:sudo sed -i 's/^#runtime = "crun"/runtime = "runc"/g' /usr/share/containers/containers.conf 2026-03-10T09:59:40.888 INFO:teuthology.task.pexec:sudo sed -i 's/runtime = "crun"/#runtime = "crun"/g' /usr/share/containers/containers.conf 2026-03-10T09:59:41.068 INFO:teuthology.orchestra.run.vm09.stdout:No match for argument: nvme-cli 2026-03-10T09:59:41.068 INFO:teuthology.orchestra.run.vm09.stderr:No packages marked for removal. 2026-03-10T09:59:41.071 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T09:59:41.071 INFO:teuthology.orchestra.run.vm09.stdout:Nothing to do. 2026-03-10T09:59:41.071 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T09:59:41.083 INFO:teuthology.orchestra.run.vm06.stdout:No match for argument: nvme-cli 2026-03-10T09:59:41.084 INFO:teuthology.orchestra.run.vm06.stderr:No packages marked for removal. 2026-03-10T09:59:41.091 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T09:59:41.091 INFO:teuthology.orchestra.run.vm06.stdout:Nothing to do. 2026-03-10T09:59:41.091 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T09:59:41.427 INFO:teuthology.orchestra.run.vm09.stdout:Last metadata expiration check: 0:02:03 ago on Tue 10 Mar 2026 09:57:38 AM UTC. 2026-03-10T09:59:41.466 INFO:teuthology.orchestra.run.vm06.stdout:Last metadata expiration check: 0:01:03 ago on Tue 10 Mar 2026 09:58:38 AM UTC. 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout:Dependencies resolved. 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout: Package Arch Version Repository Size 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout:Installing: 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout:Installing dependencies: 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout:Transaction Summary 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout:================================================================================ 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout:Install 7 Packages 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout:Total download size: 6.3 M 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout:Installed size: 24 M 2026-03-10T09:59:41.520 INFO:teuthology.orchestra.run.vm09.stdout:Downloading Packages: 2026-03-10T09:59:41.586 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T09:59:41.586 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T09:59:41.586 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T09:59:41.586 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T09:59:41.586 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli x86_64 2.16-1.el9 baseos 1.2 M 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout: nvmetcli noarch 0.8-3.el9 baseos 44 k 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout: runc x86_64 4:1.4.0-2.el9 appstream 4.0 M 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell noarch 1:1.1.30-1.el9 baseos 72 k 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout: python3-kmod x86_64 0.9-32.el9 baseos 84 k 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing noarch 2.4.7-9.el9 baseos 150 k 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout: python3-urwid x86_64 2.1.2-4.el9 baseos 837 k 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout:Install 7 Packages 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 6.3 M 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout:Installed size: 24 M 2026-03-10T09:59:41.587 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-10T09:59:41.925 INFO:teuthology.orchestra.run.vm09.stdout:(1/7): python3-configshell-1.1.30-1.el9.noarch. 398 kB/s | 72 kB 00:00 2026-03-10T09:59:41.925 INFO:teuthology.orchestra.run.vm09.stdout:(2/7): nvmetcli-0.8-3.el9.noarch.rpm 241 kB/s | 44 kB 00:00 2026-03-10T09:59:41.990 INFO:teuthology.orchestra.run.vm09.stdout:(3/7): python3-kmod-0.9-32.el9.x86_64.rpm 1.3 MB/s | 84 kB 00:00 2026-03-10T09:59:41.991 INFO:teuthology.orchestra.run.vm09.stdout:(4/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 2.2 MB/s | 150 kB 00:00 2026-03-10T09:59:42.056 INFO:teuthology.orchestra.run.vm09.stdout:(5/7): nvme-cli-2.16-1.el9.x86_64.rpm 3.7 MB/s | 1.2 MB 00:00 2026-03-10T09:59:42.094 INFO:teuthology.orchestra.run.vm09.stdout:(6/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 7.9 MB/s | 837 kB 00:00 2026-03-10T09:59:42.246 INFO:teuthology.orchestra.run.vm09.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 16 MB/s | 4.0 MB 00:00 2026-03-10T09:59:42.246 INFO:teuthology.orchestra.run.vm09.stdout:-------------------------------------------------------------------------------- 2026-03-10T09:59:42.246 INFO:teuthology.orchestra.run.vm09.stdout:Total 8.6 MB/s | 6.3 MB 00:00 2026-03-10T09:59:42.323 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction check 2026-03-10T09:59:42.332 INFO:teuthology.orchestra.run.vm09.stdout:Transaction check succeeded. 2026-03-10T09:59:42.332 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction test 2026-03-10T09:59:42.393 INFO:teuthology.orchestra.run.vm09.stdout:Transaction test succeeded. 2026-03-10T09:59:42.394 INFO:teuthology.orchestra.run.vm09.stdout:Running transaction 2026-03-10T09:59:42.557 INFO:teuthology.orchestra.run.vm09.stdout: Preparing : 1/1 2026-03-10T09:59:42.568 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-10T09:59:42.579 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-10T09:59:42.586 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-10T09:59:42.594 INFO:teuthology.orchestra.run.vm09.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-10T09:59:42.596 INFO:teuthology.orchestra.run.vm09.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-10T09:59:42.658 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-10T09:59:42.727 INFO:teuthology.orchestra.run.vm06.stdout:(1/7): nvmetcli-0.8-3.el9.noarch.rpm 320 kB/s | 44 kB 00:00 2026-03-10T09:59:42.752 INFO:teuthology.orchestra.run.vm06.stdout:(2/7): python3-configshell-1.1.30-1.el9.noarch. 443 kB/s | 72 kB 00:00 2026-03-10T09:59:42.799 INFO:teuthology.orchestra.run.vm09.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-10T09:59:42.805 INFO:teuthology.orchestra.run.vm09.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-10T09:59:42.832 INFO:teuthology.orchestra.run.vm06.stdout:(3/7): nvme-cli-2.16-1.el9.x86_64.rpm 4.8 MB/s | 1.2 MB 00:00 2026-03-10T09:59:42.833 INFO:teuthology.orchestra.run.vm06.stdout:(4/7): python3-kmod-0.9-32.el9.x86_64.rpm 796 kB/s | 84 kB 00:00 2026-03-10T09:59:42.834 INFO:teuthology.orchestra.run.vm06.stdout:(5/7): python3-pyparsing-2.4.7-9.el9.noarch.rpm 1.8 MB/s | 150 kB 00:00 2026-03-10T09:59:42.908 INFO:teuthology.orchestra.run.vm06.stdout:(6/7): python3-urwid-2.1.2-4.el9.x86_64.rpm 11 MB/s | 837 kB 00:00 2026-03-10T09:59:43.107 INFO:teuthology.orchestra.run.vm09.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-10T09:59:43.107 INFO:teuthology.orchestra.run.vm09.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T09:59:43.107 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T09:59:43.227 INFO:teuthology.orchestra.run.vm06.stdout:(7/7): runc-1.4.0-2.el9.x86_64.rpm 10 MB/s | 4.0 MB 00:00 2026-03-10T09:59:43.229 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-10T09:59:43.230 INFO:teuthology.orchestra.run.vm06.stdout:Total 3.8 MB/s | 6.3 MB 00:01 2026-03-10T09:59:43.346 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T09:59:43.356 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T09:59:43.356 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T09:59:43.439 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T09:59:43.440 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T09:59:43.565 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-10T09:59:43.565 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-10T09:59:43.565 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-10T09:59:43.565 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-10T09:59:43.565 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-10T09:59:43.565 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-10T09:59:43.626 INFO:teuthology.orchestra.run.vm09.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-10T09:59:43.627 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T09:59:43.627 INFO:teuthology.orchestra.run.vm09.stdout:Installed: 2026-03-10T09:59:43.627 INFO:teuthology.orchestra.run.vm09.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T09:59:43.627 INFO:teuthology.orchestra.run.vm09.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T09:59:43.627 INFO:teuthology.orchestra.run.vm09.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-10T09:59:43.627 INFO:teuthology.orchestra.run.vm09.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-10T09:59:43.627 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T09:59:43.627 INFO:teuthology.orchestra.run.vm09.stdout:Complete! 2026-03-10T09:59:43.647 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T09:59:43.662 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urwid-2.1.2-4.el9.x86_64 1/7 2026-03-10T09:59:43.678 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-pyparsing-2.4.7-9.el9.noarch 2/7 2026-03-10T09:59:43.685 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-10T09:59:43.696 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-10T09:59:43.701 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvmetcli-0.8-3.el9.noarch 5/7 2026-03-10T09:59:43.707 DEBUG:teuthology.parallel:result is None 2026-03-10T09:59:43.767 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvmetcli-0.8-3.el9.noarch 5/7 2026-03-10T09:59:43.946 INFO:teuthology.orchestra.run.vm06.stdout: Installing : runc-4:1.4.0-2.el9.x86_64 6/7 2026-03-10T09:59:43.954 INFO:teuthology.orchestra.run.vm06.stdout: Installing : nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-10T09:59:44.387 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: nvme-cli-2.16-1.el9.x86_64 7/7 2026-03-10T09:59:44.387 INFO:teuthology.orchestra.run.vm06.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /usr/lib/systemd/system/nvmefc-boot-connections.service. 2026-03-10T09:59:44.387 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T09:59:44.922 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvme-cli-2.16-1.el9.x86_64 1/7 2026-03-10T09:59:44.923 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : nvmetcli-0.8-3.el9.noarch 2/7 2026-03-10T09:59:44.923 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-configshell-1:1.1.30-1.el9.noarch 3/7 2026-03-10T09:59:44.923 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-kmod-0.9-32.el9.x86_64 4/7 2026-03-10T09:59:44.923 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-pyparsing-2.4.7-9.el9.noarch 5/7 2026-03-10T09:59:44.923 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urwid-2.1.2-4.el9.x86_64 6/7 2026-03-10T09:59:45.011 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : runc-4:1.4.0-2.el9.x86_64 7/7 2026-03-10T09:59:45.011 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T09:59:45.011 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-10T09:59:45.011 INFO:teuthology.orchestra.run.vm06.stdout: nvme-cli-2.16-1.el9.x86_64 nvmetcli-0.8-3.el9.noarch 2026-03-10T09:59:45.011 INFO:teuthology.orchestra.run.vm06.stdout: python3-configshell-1:1.1.30-1.el9.noarch python3-kmod-0.9-32.el9.x86_64 2026-03-10T09:59:45.011 INFO:teuthology.orchestra.run.vm06.stdout: python3-pyparsing-2.4.7-9.el9.noarch python3-urwid-2.1.2-4.el9.x86_64 2026-03-10T09:59:45.011 INFO:teuthology.orchestra.run.vm06.stdout: runc-4:1.4.0-2.el9.x86_64 2026-03-10T09:59:45.011 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T09:59:45.011 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T09:59:45.146 DEBUG:teuthology.parallel:result is None 2026-03-10T09:59:45.146 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-10T09:59:45.195 INFO:tasks.cephadm:Config: {'roleless': True, 'conf': {'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000, 'osd shutdown pgref assert': True}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'CEPHADM_DAEMON_PLACE_FAIL', 'CEPHADM_FAILED_DAEMON'], 'log-only-match': ['CEPHADM_'], 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df'} 2026-03-10T09:59:45.195 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T09:59:45.195 INFO:tasks.cephadm:Cluster fsid is ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T09:59:45.195 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-10T09:59:45.195 INFO:tasks.cephadm:No mon roles; fabricating mons 2026-03-10T09:59:45.195 INFO:tasks.cephadm:Monitor IPs: {'mon.vm06': '192.168.123.106', 'mon.vm09': '192.168.123.109'} 2026-03-10T09:59:45.195 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-10T09:59:45.195 DEBUG:teuthology.orchestra.run.vm06:> sudo hostname $(hostname -s) 2026-03-10T09:59:45.242 DEBUG:teuthology.orchestra.run.vm09:> sudo hostname $(hostname -s) 2026-03-10T09:59:45.269 INFO:tasks.cephadm:Downloading "compiled" cephadm from cachra 2026-03-10T09:59:45.269 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=centos%2F9%2Fx86_64&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T09:59:45.905 INFO:tasks.cephadm:builder_project result: [{'url': 'https://3.chacra.ceph.com/r/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'chacra_url': 'https://3.chacra.ceph.com/repos/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/flavors/default/', 'ref': 'squid', 'sha1': 'e911bdebe5c8faa3800735d1568fcdca65db60df', 'distro': 'centos', 'distro_version': '9', 'distro_codename': None, 'modified': '2026-02-25 18:55:15.146628', 'status': 'ready', 'flavor': 'default', 'project': 'ceph', 'archs': ['source', 'x86_64'], 'extra': {'version': '19.2.3-678-ge911bdeb', 'package_manager_version': '19.2.3-678.ge911bdeb', 'build_url': 'https://jenkins.ceph.com/job/ceph-dev-pipeline/3275/', 'root_build_cause': '', 'node_name': '10.20.192.26+soko16', 'job_name': 'ceph-dev-pipeline'}}] 2026-03-10T09:59:46.505 INFO:tasks.util.chacra:got chacra host 3.chacra.ceph.com, ref squid, sha1 e911bdebe5c8faa3800735d1568fcdca65db60df from https://shaman.ceph.com/api/search/?project=ceph&distros=centos%2F9%2Fx86_64&flavor=default&sha1=e911bdebe5c8faa3800735d1568fcdca65db60df 2026-03-10T09:59:46.505 INFO:tasks.cephadm:Discovered cachra url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-10T09:59:46.505 INFO:tasks.cephadm:Downloading cephadm from url: https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm 2026-03-10T09:59:46.506 DEBUG:teuthology.orchestra.run.vm06:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T09:59:47.910 INFO:teuthology.orchestra.run.vm06.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 10 09:59 /home/ubuntu/cephtest/cephadm 2026-03-10T09:59:47.911 DEBUG:teuthology.orchestra.run.vm09:> curl --silent -L https://3.chacra.ceph.com/binaries/ceph/squid/e911bdebe5c8faa3800735d1568fcdca65db60df/centos/9/x86_64/flavors/default/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-10T09:59:49.228 INFO:teuthology.orchestra.run.vm09.stdout:-rw-r--r--. 1 ubuntu ubuntu 788355 Mar 10 09:59 /home/ubuntu/cephtest/cephadm 2026-03-10T09:59:49.228 DEBUG:teuthology.orchestra.run.vm06:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T09:59:49.250 DEBUG:teuthology.orchestra.run.vm09:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-10T09:59:49.269 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df on all hosts... 2026-03-10T09:59:49.269 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T09:59:49.293 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df pull 2026-03-10T09:59:49.476 INFO:teuthology.orchestra.run.vm09.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T09:59:49.504 INFO:teuthology.orchestra.run.vm06.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T10:00:24.593 INFO:teuthology.orchestra.run.vm06.stdout:{ 2026-03-10T10:00:24.593 INFO:teuthology.orchestra.run.vm06.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T10:00:24.593 INFO:teuthology.orchestra.run.vm06.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T10:00:24.593 INFO:teuthology.orchestra.run.vm06.stdout: "repo_digests": [ 2026-03-10T10:00:24.593 INFO:teuthology.orchestra.run.vm06.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T10:00:24.593 INFO:teuthology.orchestra.run.vm06.stdout: ] 2026-03-10T10:00:24.593 INFO:teuthology.orchestra.run.vm06.stdout:} 2026-03-10T10:00:28.952 INFO:teuthology.orchestra.run.vm09.stdout:{ 2026-03-10T10:00:28.952 INFO:teuthology.orchestra.run.vm09.stdout: "ceph_version": "ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable)", 2026-03-10T10:00:28.952 INFO:teuthology.orchestra.run.vm09.stdout: "image_id": "654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c", 2026-03-10T10:00:28.952 INFO:teuthology.orchestra.run.vm09.stdout: "repo_digests": [ 2026-03-10T10:00:28.952 INFO:teuthology.orchestra.run.vm09.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc" 2026-03-10T10:00:28.952 INFO:teuthology.orchestra.run.vm09.stdout: ] 2026-03-10T10:00:28.952 INFO:teuthology.orchestra.run.vm09.stdout:} 2026-03-10T10:00:28.974 DEBUG:teuthology.orchestra.run.vm06:> sudo mkdir -p /etc/ceph 2026-03-10T10:00:29.008 DEBUG:teuthology.orchestra.run.vm09:> sudo mkdir -p /etc/ceph 2026-03-10T10:00:29.038 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 777 /etc/ceph 2026-03-10T10:00:29.077 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 777 /etc/ceph 2026-03-10T10:00:29.106 INFO:tasks.cephadm:Writing seed config... 2026-03-10T10:00:29.107 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-10T10:00:29.107 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-10T10:00:29.107 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-10T10:00:29.107 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-10T10:00:29.107 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-10T10:00:29.107 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-10T10:00:29.107 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-10T10:00:29.107 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-10T10:00:29.107 INFO:tasks.cephadm: override: [osd] osd shutdown pgref assert = True 2026-03-10T10:00:29.107 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T10:00:29.107 DEBUG:teuthology.orchestra.run.vm06:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-10T10:00:29.138 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = ddb3da90-1c67-11f1-bef5-37be0d23c456 [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = True bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-10T10:00:29.138 DEBUG:teuthology.orchestra.run.vm06:mon.vm06> sudo journalctl -f -n 0 -u ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mon.vm06.service 2026-03-10T10:00:29.180 INFO:tasks.cephadm:Bootstrapping... 2026-03-10T10:00:29.180 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df -v bootstrap --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-ip 192.168.123.106 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:00:29.383 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-10T10:00:29.383 INFO:teuthology.orchestra.run.vm06.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df', '-v', 'bootstrap', '--fsid', 'ddb3da90-1c67-11f1-bef5-37be0d23c456', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-ip', '192.168.123.106', '--skip-admin-label'] 2026-03-10T10:00:29.383 INFO:teuthology.orchestra.run.vm06.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-10T10:00:29.383 INFO:teuthology.orchestra.run.vm06.stdout:Verifying podman|docker is present... 2026-03-10T10:00:29.404 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stdout 5.8.0 2026-03-10T10:00:29.404 INFO:teuthology.orchestra.run.vm06.stdout:Verifying lvm2 is present... 2026-03-10T10:00:29.404 INFO:teuthology.orchestra.run.vm06.stdout:Verifying time synchronization is in place... 2026-03-10T10:00:29.413 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T10:00:29.413 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T10:00:29.424 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T10:00:29.425 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-10T10:00:29.432 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout enabled 2026-03-10T10:00:29.440 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout active 2026-03-10T10:00:29.441 INFO:teuthology.orchestra.run.vm06.stdout:Unit chronyd.service is enabled and running 2026-03-10T10:00:29.441 INFO:teuthology.orchestra.run.vm06.stdout:Repeating the final host check... 2026-03-10T10:00:29.468 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stdout 5.8.0 2026-03-10T10:00:29.469 INFO:teuthology.orchestra.run.vm06.stdout:podman (/bin/podman) version 5.8.0 is present 2026-03-10T10:00:29.469 INFO:teuthology.orchestra.run.vm06.stdout:systemctl is present 2026-03-10T10:00:29.469 INFO:teuthology.orchestra.run.vm06.stdout:lvcreate is present 2026-03-10T10:00:29.481 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-10T10:00:29.481 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-10T10:00:29.489 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-10T10:00:29.489 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout inactive 2026-03-10T10:00:29.499 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout enabled 2026-03-10T10:00:29.506 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stdout active 2026-03-10T10:00:29.506 INFO:teuthology.orchestra.run.vm06.stdout:Unit chronyd.service is enabled and running 2026-03-10T10:00:29.506 INFO:teuthology.orchestra.run.vm06.stdout:Host looks OK 2026-03-10T10:00:29.507 INFO:teuthology.orchestra.run.vm06.stdout:Cluster fsid: ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:00:29.507 INFO:teuthology.orchestra.run.vm06.stdout:Acquiring lock 139991030812480 on /run/cephadm/ddb3da90-1c67-11f1-bef5-37be0d23c456.lock 2026-03-10T10:00:29.507 INFO:teuthology.orchestra.run.vm06.stdout:Lock 139991030812480 acquired on /run/cephadm/ddb3da90-1c67-11f1-bef5-37be0d23c456.lock 2026-03-10T10:00:29.508 INFO:teuthology.orchestra.run.vm06.stdout:Verifying IP 192.168.123.106 port 3300 ... 2026-03-10T10:00:29.508 INFO:teuthology.orchestra.run.vm06.stdout:Verifying IP 192.168.123.106 port 6789 ... 2026-03-10T10:00:29.508 INFO:teuthology.orchestra.run.vm06.stdout:Base mon IP(s) is [192.168.123.106:3300, 192.168.123.106:6789], mon addrv is [v2:192.168.123.106:3300,v1:192.168.123.106:6789] 2026-03-10T10:00:29.513 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.106 metric 100 2026-03-10T10:00:29.513 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout 192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.106 metric 100 2026-03-10T10:00:29.517 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-10T10:00:29.517 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout fe80::/64 dev eth0 proto kernel metric 1024 pref medium 2026-03-10T10:00:29.521 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-10T10:00:29.521 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-10T10:00:29.521 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T10:00:29.521 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout 2: eth0: mtu 1500 state UP qlen 1000 2026-03-10T10:00:29.522 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout inet6 fe80::5055:ff:fe00:6/64 scope link noprefixroute 2026-03-10T10:00:29.522 INFO:teuthology.orchestra.run.vm06.stdout:/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-10T10:00:29.522 INFO:teuthology.orchestra.run.vm06.stdout:Mon IP `192.168.123.106` is in CIDR network `192.168.123.0/24` 2026-03-10T10:00:29.522 INFO:teuthology.orchestra.run.vm06.stdout:Mon IP `192.168.123.106` is in CIDR network `192.168.123.0/24` 2026-03-10T10:00:29.522 INFO:teuthology.orchestra.run.vm06.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24'] 2026-03-10T10:00:29.523 INFO:teuthology.orchestra.run.vm06.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-10T10:00:29.523 INFO:teuthology.orchestra.run.vm06.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T10:00:31.214 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stdout 654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-10T10:00:31.214 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Trying to pull quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df... 2026-03-10T10:00:31.214 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Getting image source signatures 2026-03-10T10:00:31.214 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Copying blob sha256:1752b8d01aa0dd33bbe0ab24e8316174c94fbdcd5d26252e2680bba0624747a7 2026-03-10T10:00:31.214 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Copying blob sha256:8e380faede39ebd4286247457b408d979ab568aafd8389c42ec304b8cfba4e92 2026-03-10T10:00:31.214 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Copying config sha256:654f31e6858eb235bbece362255b685a945f2b6a367e2b88c4930c984fbb214c 2026-03-10T10:00:31.214 INFO:teuthology.orchestra.run.vm06.stdout:/bin/podman: stderr Writing manifest to image destination 2026-03-10T10:00:31.504 INFO:teuthology.orchestra.run.vm06.stdout:ceph: stdout ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T10:00:31.504 INFO:teuthology.orchestra.run.vm06.stdout:Ceph version: ceph version 19.2.3-678-ge911bdeb (e911bdebe5c8faa3800735d1568fcdca65db60df) squid (stable) 2026-03-10T10:00:31.504 INFO:teuthology.orchestra.run.vm06.stdout:Extracting ceph user uid/gid from container image... 2026-03-10T10:00:31.747 INFO:teuthology.orchestra.run.vm06.stdout:stat: stdout 167 167 2026-03-10T10:00:31.747 INFO:teuthology.orchestra.run.vm06.stdout:Creating initial keys... 2026-03-10T10:00:31.995 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-authtool: stdout AQC/669pHbHpMhAAVs32C5L3SV1MkY+pL91jQQ== 2026-03-10T10:00:32.242 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-authtool: stdout AQDA669pGK4WBhAA+qIX0XxJeiQ6mx0dtiKINw== 2026-03-10T10:00:32.475 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph-authtool: stdout AQDA669pWESQFBAAIPVJPol5vfJPt7i5YLZLfA== 2026-03-10T10:00:32.476 INFO:teuthology.orchestra.run.vm06.stdout:Creating initial monmap... 2026-03-10T10:00:32.724 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T10:00:32.724 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = quincy 2026-03-10T10:00:32.724 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:00:32.724 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T10:00:32.724 INFO:teuthology.orchestra.run.vm06.stdout:monmaptool for vm06 [v2:192.168.123.106:3300,v1:192.168.123.106:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-10T10:00:32.724 INFO:teuthology.orchestra.run.vm06.stdout:setting min_mon_release = quincy 2026-03-10T10:00:32.724 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: set fsid to ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:00:32.724 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-10T10:00:32.724 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:00:32.724 INFO:teuthology.orchestra.run.vm06.stdout:Creating mon... 2026-03-10T10:00:32.975 INFO:teuthology.orchestra.run.vm06.stdout:create mon.vm06 on 2026-03-10T10:00:33.298 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-10T10:00:33.454 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456.target → /etc/systemd/system/ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456.target. 2026-03-10T10:00:33.454 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456.target → /etc/systemd/system/ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456.target. 2026-03-10T10:00:33.629 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mon.vm06 2026-03-10T10:00:33.629 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to reset failed state of unit ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mon.vm06.service: Unit ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mon.vm06.service not loaded. 2026-03-10T10:00:33.788 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456.target.wants/ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mon.vm06.service → /etc/systemd/system/ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@.service. 2026-03-10T10:00:34.019 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-10T10:00:34.019 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T10:00:34.019 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mon to start... 2026-03-10T10:00:34.019 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mon... 2026-03-10T10:00:34.361 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout cluster: 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout id: ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout services: 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum vm06 (age 0.169937s) 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout data: 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout pgs: 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:mon is available 2026-03-10T10:00:34.362 INFO:teuthology.orchestra.run.vm06.stdout:Assimilating anything we can from ceph.conf... 2026-03-10T10:00:34.690 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout fsid = ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.106:3300,v1:192.168.123.106:6789] 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T10:00:34.691 INFO:teuthology.orchestra.run.vm06.stdout:Generating new minimal ceph.conf... 2026-03-10T10:00:35.011 INFO:teuthology.orchestra.run.vm06.stdout:Restarting the monitor... 2026-03-10T10:00:35.652 INFO:teuthology.orchestra.run.vm06.stdout:Setting public_network to 192.168.123.0/24 in mon config section 2026-03-10T10:00:36.001 INFO:teuthology.orchestra.run.vm06.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-10T10:00:36.001 INFO:teuthology.orchestra.run.vm06.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:00:36.001 INFO:teuthology.orchestra.run.vm06.stdout:Creating mgr... 2026-03-10T10:00:36.002 INFO:teuthology.orchestra.run.vm06.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-10T10:00:36.002 INFO:teuthology.orchestra.run.vm06.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-10T10:00:36.002 INFO:teuthology.orchestra.run.vm06.stdout:Verifying port 0.0.0.0:8443 ... 2026-03-10T10:00:36.164 INFO:teuthology.orchestra.run.vm06.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mgr.vm06.dkjjvn 2026-03-10T10:00:36.164 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Failed to reset failed state of unit ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mgr.vm06.dkjjvn.service: Unit ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mgr.vm06.dkjjvn.service not loaded. 2026-03-10T10:00:36.315 INFO:teuthology.orchestra.run.vm06.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456.target.wants/ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mgr.vm06.dkjjvn.service → /etc/systemd/system/ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@.service. 2026-03-10T10:00:36.549 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-10T10:00:36.549 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to enable service . firewalld.service is not available 2026-03-10T10:00:36.549 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-10T10:00:36.549 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to open ports <[9283, 8765, 8443]>. firewalld.service is not available 2026-03-10T10:00:36.550 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr to start... 2026-03-10T10:00:36.550 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr... 2026-03-10T10:00:36.924 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:36.924 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:00:36.924 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsid": "ddb3da90-1c67-11f1-bef5-37be0d23c456", 2026-03-10T10:00:36.924 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T10:00:36.924 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T10:00:36.924 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T10:00:36.924 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T10:00:36.924 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 0 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "vm06" 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_age": 1, 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T10:00:36.925 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T10:00:34:057451+0000", 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T10:00:34.058928+0000", 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:00:36.926 INFO:teuthology.orchestra.run.vm06.stdout:mgr not available, waiting (1/15)... 2026-03-10T10:00:39.295 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:39.295 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsid": "ddb3da90-1c67-11f1-bef5-37be0d23c456", 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 0 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "vm06" 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_age": 3, 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T10:00:39.296 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T10:00:34:057451+0000", 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T10:00:34.058928+0000", 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:00:39.297 INFO:teuthology.orchestra.run.vm06.stdout:mgr not available, waiting (2/15)... 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsid": "ddb3da90-1c67-11f1-bef5-37be0d23c456", 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "health": { 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 0 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "vm06" 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "quorum_age": 6, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "squid", 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-10T10:00:41.961 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "btime": "2026-03-10T10:00:34:057451+0000", 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-10T10:00:41.962 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "nfs", 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "restful" 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ], 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "modified": "2026-03-10T10:00:34.058928+0000", 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout }, 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:00:41.963 INFO:teuthology.orchestra.run.vm06.stdout:mgr is available 2026-03-10T10:00:42.341 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:42.341 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [global] 2026-03-10T10:00:42.341 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout fsid = ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:00:42.341 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-10T10:00:42.341 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.106:3300,v1:192.168.123.106:6789] 2026-03-10T10:00:42.341 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-10T10:00:42.341 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-10T10:00:42.341 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-10T10:00:42.341 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-10T10:00:42.341 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:42.342 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-10T10:00:42.342 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-10T10:00:42.342 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 2026-03-10T10:00:42.342 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout [osd] 2026-03-10T10:00:42.342 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-10T10:00:42.342 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-10T10:00:42.342 INFO:teuthology.orchestra.run.vm06.stdout:Enabling cephadm module... 2026-03-10T10:00:43.305 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:00:43.305 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-10T10:00:43.305 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T10:00:43.305 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "active_name": "vm06.dkjjvn", 2026-03-10T10:00:43.305 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T10:00:43.305 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:00:43.305 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for the mgr to restart... 2026-03-10T10:00:43.305 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr epoch 5... 2026-03-10T10:00:47.134 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:00:47.134 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-10T10:00:47.134 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T10:00:47.134 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:00:47.134 INFO:teuthology.orchestra.run.vm06.stdout:mgr epoch 5 is available 2026-03-10T10:00:47.134 INFO:teuthology.orchestra.run.vm06.stdout:Setting orchestrator backend to cephadm... 2026-03-10T10:00:47.954 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-10T10:00:47.954 INFO:teuthology.orchestra.run.vm06.stdout:Generating ssh key... 2026-03-10T10:00:48.778 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKkyVwlnpqSJV8XwdAvlZM60USJNshumsxR5lKfgGCZC ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:00:48.778 INFO:teuthology.orchestra.run.vm06.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-10T10:00:48.778 INFO:teuthology.orchestra.run.vm06.stdout:Adding key to root@localhost authorized_keys... 2026-03-10T10:00:48.778 INFO:teuthology.orchestra.run.vm06.stdout:Adding host vm06... 2026-03-10T10:00:50.805 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Added host 'vm06' with addr '192.168.123.106' 2026-03-10T10:00:50.805 INFO:teuthology.orchestra.run.vm06.stdout:Deploying mon service with default placement... 2026-03-10T10:00:51.274 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-10T10:00:51.274 INFO:teuthology.orchestra.run.vm06.stdout:Deploying mgr service with default placement... 2026-03-10T10:00:51.681 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-10T10:00:51.681 INFO:teuthology.orchestra.run.vm06.stdout:Deploying crash service with default placement... 2026-03-10T10:00:52.071 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled crash update... 2026-03-10T10:00:52.071 INFO:teuthology.orchestra.run.vm06.stdout:Deploying ceph-exporter service with default placement... 2026-03-10T10:00:52.528 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled ceph-exporter update... 2026-03-10T10:00:52.528 INFO:teuthology.orchestra.run.vm06.stdout:Deploying prometheus service with default placement... 2026-03-10T10:00:53.066 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled prometheus update... 2026-03-10T10:00:53.066 INFO:teuthology.orchestra.run.vm06.stdout:Deploying grafana service with default placement... 2026-03-10T10:00:53.529 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled grafana update... 2026-03-10T10:00:53.529 INFO:teuthology.orchestra.run.vm06.stdout:Deploying node-exporter service with default placement... 2026-03-10T10:00:54.037 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled node-exporter update... 2026-03-10T10:00:54.037 INFO:teuthology.orchestra.run.vm06.stdout:Deploying alertmanager service with default placement... 2026-03-10T10:00:54.420 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Scheduled alertmanager update... 2026-03-10T10:00:55.179 INFO:teuthology.orchestra.run.vm06.stdout:Enabling the dashboard module... 2026-03-10T10:00:56.039 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:55 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1515328673' entity='client.admin' cmd=[{"prefix": "mgr module enable", "module": "dashboard"}]: dispatch 2026-03-10T10:00:56.641 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:00:56.641 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-10T10:00:56.641 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-10T10:00:56.641 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "active_name": "vm06.dkjjvn", 2026-03-10T10:00:56.641 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-10T10:00:56.641 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:00:56.641 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for the mgr to restart... 2026-03-10T10:00:56.641 INFO:teuthology.orchestra.run.vm06.stdout:Waiting for mgr epoch 9... 2026-03-10T10:00:57.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:57 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1515328673' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-10T10:00:57.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:57 vm06 ceph-mon[53281]: mgrmap e9: vm06.dkjjvn(active, since 10s) 2026-03-10T10:00:57.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:57 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/2557393409' entity='client.admin' cmd=[{"prefix": "mgr stat"}]: dispatch 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: Active manager daemon vm06.dkjjvn restarted 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: Activating manager daemon vm06.dkjjvn 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: osdmap e3: 0 total, 0 up, 0 in 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: mgrmap e10: vm06.dkjjvn(active, starting, since 0.0147854s) 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm06"}]: dispatch 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mgr metadata", "who": "vm06.dkjjvn", "id": "vm06.dkjjvn"}]: dispatch 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: Manager daemon vm06.dkjjvn is now available 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm06.dkjjvn/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm06.dkjjvn/trash_purge_schedule"}]: dispatch 2026-03-10T10:00:59.507 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:00:59 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:01:00.272 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout { 2026-03-10T10:01:00.272 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-10T10:01:00.272 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-10T10:01:00.272 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout } 2026-03-10T10:01:00.272 INFO:teuthology.orchestra.run.vm06.stdout:mgr epoch 9 is available 2026-03-10T10:01:00.272 INFO:teuthology.orchestra.run.vm06.stdout:Generating a dashboard self-signed certificate... 2026-03-10T10:01:00.658 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:00 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:00.658 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:00 vm06 ceph-mon[53281]: [10/Mar/2026:10:00:59] ENGINE Bus STARTING 2026-03-10T10:01:00.658 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:00 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:00.658 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:00 vm06 ceph-mon[53281]: [10/Mar/2026:10:01:00] ENGINE Serving on https://192.168.123.106:7150 2026-03-10T10:01:00.658 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:00 vm06 ceph-mon[53281]: [10/Mar/2026:10:01:00] ENGINE Client ('192.168.123.106', 46474) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:01:00.658 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:00 vm06 ceph-mon[53281]: mgrmap e11: vm06.dkjjvn(active, since 1.01817s) 2026-03-10T10:01:00.658 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:00 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:00.658 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:00 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:00.714 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout Self-signed certificate created 2026-03-10T10:01:00.714 INFO:teuthology.orchestra.run.vm06.stdout:Creating initial admin user... 2026-03-10T10:01:01.219 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$aAg0lJutI7gi8r8eNwDU0umoieAHgpbiVDWAmLFBXuwz5coFoC/be", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1773136861, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-10T10:01:01.220 INFO:teuthology.orchestra.run.vm06.stdout:Fetching dashboard port number... 2026-03-10T10:01:01.580 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stdout 8443 2026-03-10T10:01:01.580 INFO:teuthology.orchestra.run.vm06.stdout:firewalld does not appear to be present 2026-03-10T10:01:01.580 INFO:teuthology.orchestra.run.vm06.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-10T10:01:01.582 INFO:teuthology.orchestra.run.vm06.stdout:Ceph Dashboard is now available at: 2026-03-10T10:01:01.582 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:01.582 INFO:teuthology.orchestra.run.vm06.stdout: URL: https://vm06.local:8443/ 2026-03-10T10:01:01.582 INFO:teuthology.orchestra.run.vm06.stdout: User: admin 2026-03-10T10:01:01.582 INFO:teuthology.orchestra.run.vm06.stdout: Password: j9ckhpo703 2026-03-10T10:01:01.582 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:01.582 INFO:teuthology.orchestra.run.vm06.stdout:Saving cluster configuration to /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/config directory 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout:Or, if you are only running a single cluster on this host: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: ceph telemetry on 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout:For more information see: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:01.983 INFO:teuthology.orchestra.run.vm06.stdout:Bootstrap complete. 2026-03-10T10:01:02.012 INFO:tasks.cephadm:Fetching config... 2026-03-10T10:01:02.012 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T10:01:02.012 DEBUG:teuthology.orchestra.run.vm06:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-10T10:01:02.028 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-10T10:01:02.028 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T10:01:02.028 DEBUG:teuthology.orchestra.run.vm06:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-10T10:01:02.088 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-10T10:01:02.088 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T10:01:02.088 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/keyring of=/dev/stdout 2026-03-10T10:01:02.153 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-10T10:01:02.153 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T10:01:02.153 DEBUG:teuthology.orchestra.run.vm06:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-10T10:01:02.208 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-10T10:01:02.208 DEBUG:teuthology.orchestra.run.vm06:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKkyVwlnpqSJV8XwdAvlZM60USJNshumsxR5lKfgGCZC ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T10:01:02.273 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:02 vm06 ceph-mon[53281]: from='client.14166 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-10T10:01:02.273 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:02 vm06 ceph-mon[53281]: from='client.14166 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-10T10:01:02.273 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:02 vm06 ceph-mon[53281]: [10/Mar/2026:10:01:00] ENGINE Serving on http://192.168.123.106:8765 2026-03-10T10:01:02.273 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:02 vm06 ceph-mon[53281]: [10/Mar/2026:10:01:00] ENGINE Bus STARTED 2026-03-10T10:01:02.273 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:02 vm06 ceph-mon[53281]: from='client.14174 -' entity='client.admin' cmd=[{"prefix": "dashboard create-self-signed-cert", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:01:02.273 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:02 vm06 ceph-mon[53281]: from='client.14176 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:01:02.273 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:02 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:02.273 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:02 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3679910211' entity='client.admin' cmd=[{"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"}]: dispatch 2026-03-10T10:01:02.273 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:02 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/2114646936' entity='client.admin' 2026-03-10T10:01:02.287 INFO:teuthology.orchestra.run.vm06.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKkyVwlnpqSJV8XwdAvlZM60USJNshumsxR5lKfgGCZC ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:01:02.321 DEBUG:teuthology.orchestra.run.vm09:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKkyVwlnpqSJV8XwdAvlZM60USJNshumsxR5lKfgGCZC ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-10T10:01:02.352 INFO:teuthology.orchestra.run.vm09.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIKkyVwlnpqSJV8XwdAvlZM60USJNshumsxR5lKfgGCZC ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:01:02.361 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-10T10:01:02.533 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:03.070 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-10T10:01:03.070 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-10T10:01:03.174 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:03 vm06 ceph-mon[53281]: mgrmap e12: vm06.dkjjvn(active, since 2s) 2026-03-10T10:01:03.174 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:03 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1989095483' entity='client.admin' 2026-03-10T10:01:03.346 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:03.763 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm09 2026-03-10T10:01:03.764 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:01:03.764 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.conf 2026-03-10T10:01:03.778 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:01:03.778 DEBUG:teuthology.orchestra.run.vm09:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:01:03.832 INFO:tasks.cephadm:Adding host vm09 to orchestrator... 2026-03-10T10:01:03.832 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch host add vm09 2026-03-10T10:01:04.159 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:04.259 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:04 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:04.259 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:04 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:04.259 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:04 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:01:04.259 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:04 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:04.259 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:04 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm06", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T10:01:04.259 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:04 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm06", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-10T10:01:04.259 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:04 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:04.259 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:04 vm06 ceph-mon[53281]: Deploying daemon ceph-exporter.vm06 on vm06 2026-03-10T10:01:04.259 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:04 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:05.360 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:05 vm06 ceph-mon[53281]: from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:01:05.360 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:05 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:05.360 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:05 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:05.360 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:05 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:05.360 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:05 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:05.360 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:05 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm06", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T10:01:05.360 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:05 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm06", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-10T10:01:05.360 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:05 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:05.966 INFO:teuthology.orchestra.run.vm06.stdout:Added host 'vm09' with addr '192.168.123.109' 2026-03-10T10:01:06.203 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch host ls --format=json 2026-03-10T10:01:06.254 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:06 vm06 ceph-mon[53281]: from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:01:06.254 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:06 vm06 ceph-mon[53281]: Deploying cephadm binary to vm09 2026-03-10T10:01:06.254 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:06 vm06 ceph-mon[53281]: Deploying daemon crash.vm06 on vm06 2026-03-10T10:01:06.254 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:06 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:06.254 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:06 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:06.254 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:06 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:06.254 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:06 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:06.254 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:06 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:06.466 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:06.695 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:06.695 INFO:teuthology.orchestra.run.vm06.stdout:[{"addr": "192.168.123.106", "hostname": "vm06", "labels": [], "status": ""}, {"addr": "192.168.123.109", "hostname": "vm09", "labels": [], "status": ""}] 2026-03-10T10:01:06.860 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-10T10:01:06.861 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd crush tunables default 2026-03-10T10:01:07.014 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:07.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:07 vm06 ceph-mon[53281]: Deploying daemon node-exporter.vm06 on vm06 2026-03-10T10:01:07.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:07 vm06 ceph-mon[53281]: Added host vm09 2026-03-10T10:01:07.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:07 vm06 ceph-mon[53281]: mgrmap e13: vm06.dkjjvn(active, since 7s) 2026-03-10T10:01:08.118 INFO:teuthology.orchestra.run.vm06.stderr:adjusted tunables profile to default 2026-03-10T10:01:08.267 INFO:tasks.cephadm:Adding mon.vm06 on vm06 2026-03-10T10:01:08.267 INFO:tasks.cephadm:Adding mon.vm09 on vm09 2026-03-10T10:01:08.267 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch apply mon '2;vm06:192.168.123.106=vm06;vm09:192.168.123.109=vm09' 2026-03-10T10:01:08.434 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:08.470 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:08.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:08 vm06 ceph-mon[53281]: from='client.14189 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:01:08.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:08 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/183256898' entity='client.admin' cmd=[{"prefix": "osd crush tunables", "profile": "default"}]: dispatch 2026-03-10T10:01:08.721 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled mon update... 2026-03-10T10:01:08.867 DEBUG:teuthology.orchestra.run.vm09:mon.vm09> sudo journalctl -f -n 0 -u ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mon.vm09.service 2026-03-10T10:01:08.868 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:08.868 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:09.048 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:09.079 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:09.315 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:09.315 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:09.315 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:09.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:09 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/183256898' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-10T10:01:09.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:09 vm06 ceph-mon[53281]: osdmap e4: 0 total, 0 up, 0 in 2026-03-10T10:01:09.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:09 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:09.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:09 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:09.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:09 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:09.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:09 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:09.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:09 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:10.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:10 vm06 ceph-mon[53281]: Deploying daemon alertmanager.vm06 on vm06 2026-03-10T10:01:10.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:10 vm06 ceph-mon[53281]: from='client.14193 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "2;vm06:192.168.123.106=vm06;vm09:192.168.123.109=vm09", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:01:10.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:10 vm06 ceph-mon[53281]: Saving service mon spec with placement vm06:192.168.123.106=vm06;vm09:192.168.123.109=vm09;count:2 2026-03-10T10:01:10.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:10 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:10.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:10 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/212192800' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:10.477 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:10.477 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:10.630 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:10.664 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:10.926 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:10.926 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:10.926 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:11.390 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:11 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/3853349338' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:12.090 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:12.090 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:12.246 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:12.281 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:12.540 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:12.540 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:12.540 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: Regenerating cephadm self-signed grafana TLS certificates 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-ssl-verify", "value": "false"}]: dispatch 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:13.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:13 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/602039258' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:13.686 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:13.686 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:13.845 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:13.876 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:14.138 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:14.138 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:14.138 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:14.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:14 vm06 ceph-mon[53281]: Deploying daemon grafana.vm06 on vm06 2026-03-10T10:01:15.290 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:15.290 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:15.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:15 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/1921316759' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:15.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:15 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:15.461 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:15.499 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:15.765 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:15.765 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:15.765 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:16.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:16 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/206539989' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:16.925 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:16.925 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:17.104 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:17.146 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:17.447 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:17.447 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:17.447 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:17.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:17 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/1095787878' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:18.610 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:18.611 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:18.784 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:18.824 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:19.106 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:19.106 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:19.106 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:20.267 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:20.267 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:20 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:20 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:20 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:20 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:20 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:20 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:20 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:20 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:20 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/3341362590' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:20 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:20.434 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:20.465 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:20.734 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:20.734 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:20.734 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:21.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:21 vm06 ceph-mon[53281]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:21.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:21 vm06 ceph-mon[53281]: Deploying daemon prometheus.vm06 on vm06 2026-03-10T10:01:21.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:21 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/1657034975' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:22.080 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:22.080 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:22.260 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:22.300 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:22.583 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:22.583 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:22.583 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:23 vm06 ceph-mon[53281]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:23 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/2667664526' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:23.760 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:23.761 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:23.925 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:23.962 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:24.237 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:24.237 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:24.237 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:25.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:25 vm06 ceph-mon[53281]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:25.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:25 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/1808657296' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:25.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:25 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:25.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:25 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:25.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:25 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:25.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:25 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mgr module enable", "module": "prometheus"}]: dispatch 2026-03-10T10:01:25.387 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:25.387 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:25.556 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:25.595 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:25.874 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:25.874 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:25.875 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:26.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:26 vm06 ceph-mon[53281]: from='mgr.14162 192.168.123.106:0/80334525' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "mgr module enable", "module": "prometheus"}]': finished 2026-03-10T10:01:26.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:26 vm06 ceph-mon[53281]: mgrmap e14: vm06.dkjjvn(active, since 26s) 2026-03-10T10:01:26.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:26 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/3167958835' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:27.046 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:27.046 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:27.221 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:27.261 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:27.553 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:27.554 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:27.554 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:27.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:27 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/2755939167' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:28.716 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:28.716 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:28.882 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:28.918 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:29.375 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:29.375 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:29.375 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: Active manager daemon vm06.dkjjvn restarted 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: Activating manager daemon vm06.dkjjvn 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: osdmap e5: 0 total, 0 up, 0 in 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: mgrmap e15: vm06.dkjjvn(active, starting, since 0.298702s) 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm06"}]: dispatch 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mgr metadata", "who": "vm06.dkjjvn", "id": "vm06.dkjjvn"}]: dispatch 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mds metadata"}]: dispatch 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata"}]: dispatch 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata"}]: dispatch 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: Manager daemon vm06.dkjjvn is now available 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:01:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:29 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:01:30.277 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:30 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm06.dkjjvn/mirror_snapshot_schedule"}]: dispatch 2026-03-10T10:01:30.277 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:30 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/vm06.dkjjvn/trash_purge_schedule"}]: dispatch 2026-03-10T10:01:30.277 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:30 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/2684384130' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:30.277 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:30 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:30.277 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:30 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:30.277 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:30 vm06 ceph-mon[53281]: mgrmap e16: vm06.dkjjvn(active, since 1.30262s) 2026-03-10T10:01:30.556 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:30.556 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:31.079 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:31.121 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /etc/ceph/ceph.conf 2026-03-10T10:01:31.433 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:31.433 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:31.433 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: [10/Mar/2026:10:01:30] ENGINE Bus STARTING 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: [10/Mar/2026:10:01:30] ENGINE Serving on https://192.168.123.106:7150 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: [10/Mar/2026:10:01:30] ENGINE Client ('192.168.123.106', 44010) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: [10/Mar/2026:10:01:30] ENGINE Serving on http://192.168.123.106:8765 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: [10/Mar/2026:10:01:30] ENGINE Bus STARTED 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:01:31.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:31 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/405422781' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:32.615 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:32.616 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:32.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:32 vm06 ceph-mon[53281]: mgrmap e17: vm06.dkjjvn(active, since 2s) 2026-03-10T10:01:32.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:32 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:32.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:32 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:32.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:32 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:32.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:32 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:32.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:32 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:01:32.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:32 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:32.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:32 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:01:32.830 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/config/ceph.conf 2026-03-10T10:01:33.145 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:33.145 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:33.145 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: Updating vm09:/var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/config/ceph.conf 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: Updating vm06:/var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/config/ceph.conf 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: Updating vm09:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: Updating vm06:/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: Updating vm09:/var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/config/ceph.client.admin.keyring 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]': finished 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:33 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/220446440' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:34.343 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:34.343 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:34.702 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/config/ceph.conf 2026-03-10T10:01:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:34 vm06 ceph-mon[53281]: Updating vm06:/var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/config/ceph.client.admin.keyring 2026-03-10T10:01:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:34 vm06 ceph-mon[53281]: Deploying daemon ceph-exporter.vm09 on vm09 2026-03-10T10:01:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:34 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:34 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:34 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:34 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:34 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T10:01:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:34 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]': finished 2026-03-10T10:01:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:34 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:35.249 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:35.249 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:35.249 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:35.817 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:35 vm06 ceph-mon[53281]: Deploying daemon crash.vm09 on vm09 2026-03-10T10:01:35.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:35 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/644480581' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:35.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:35 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:35.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:35 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:35.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:35 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:35.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:35 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:36.412 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:36.412 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:36.587 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/config/ceph.conf 2026-03-10T10:01:36.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:36 vm06 ceph-mon[53281]: Deploying daemon node-exporter.vm09 on vm09 2026-03-10T10:01:36.874 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:36.874 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:36.874 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:37.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:37 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/3563370013' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:38.044 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:38.045 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:38.251 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/config/ceph.conf 2026-03-10T10:01:38.593 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:38.593 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:38.594 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:39.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:39.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:39.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:39.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:39.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.jzyfcu", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:01:39.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "mgr.vm09.jzyfcu", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]': finished 2026-03-10T10:01:39.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:01:39.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:39.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: Deploying daemon mgr.vm09.jzyfcu on vm09 2026-03-10T10:01:39.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/1251323944' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:39.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:39.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:39.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:39 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:39.854 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:39.854 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:40.088 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm09/config 2026-03-10T10:01:40.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:40 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:40.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:40 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:40.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:40 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:01:40.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:40 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:40.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:40 vm06 ceph-mon[53281]: Deploying daemon mon.vm09 on vm09 2026-03-10T10:01:40.666 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:40.666 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":1,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:00:32.599004Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-10T10:01:40.666 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 1 2026-03-10T10:01:41.867 INFO:tasks.cephadm:Waiting for 2 mons in monmap... 2026-03-10T10:01:41.868 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mon dump -f json 2026-03-10T10:01:42.067 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm09/config 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm06"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: mon.vm06 calling monitor election 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.? 192.168.123.109:0/1658576133' entity='mgr.vm09.jzyfcu' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.jzyfcu/crt"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: mon.vm09 calling monitor election 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: mon.vm06 is new leader, mons vm06,vm09 in quorum (ranks 0,1) 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: monmap epoch 2 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: last_changed 2026-03-10T10:01:40.866606+0000 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: created 2026-03-10T10:00:32.599004+0000 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: min_mon_release 19 (squid) 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: election_strategy: 1 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: 0: [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] mon.vm06 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: 1: [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] mon.vm09 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: fsmap 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: osdmap e5: 0 total, 0 up, 0 in 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: mgrmap e17: vm06.dkjjvn(active, since 17s) 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: overall HEALTH_OK 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: Standby manager daemon vm09.jzyfcu started 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.? 192.168.123.109:0/1658576133' entity='mgr.vm09.jzyfcu' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/crt"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.? 192.168.123.109:0/1658576133' entity='mgr.vm09.jzyfcu' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/vm09.jzyfcu/key"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.? 192.168.123.109:0/1658576133' entity='mgr.vm09.jzyfcu' cmd=[{"prefix": "config-key get", "key": "mgr/dashboard/key"}]: dispatch 2026-03-10T10:01:46.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:45 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:46.800 INFO:teuthology.orchestra.run.vm09.stdout: 2026-03-10T10:01:46.800 INFO:teuthology.orchestra.run.vm09.stdout:{"epoch":2,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","modified":"2026-03-10T10:01:40.866606Z","created":"2026-03-10T10:00:32.599004Z","min_mon_release":19,"min_mon_release_name":"squid","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid"],"optional":[]},"mons":[{"rank":0,"name":"vm06","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:3300","nonce":0},{"type":"v1","addr":"192.168.123.106:6789","nonce":0}]},"addr":"192.168.123.106:6789/0","public_addr":"192.168.123.106:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"vm09","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:3300","nonce":0},{"type":"v1","addr":"192.168.123.109:6789","nonce":0}]},"addr":"192.168.123.109:6789/0","public_addr":"192.168.123.109:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-10T10:01:46.800 INFO:teuthology.orchestra.run.vm09.stderr:dumped monmap epoch 2 2026-03-10T10:01:46.949 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-10T10:01:46.949 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph config generate-minimal-conf 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: mgrmap e18: vm06.dkjjvn(active, since 17s), standbys: vm09.jzyfcu 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mgr metadata", "who": "vm09.jzyfcu", "id": "vm09.jzyfcu"}]: dispatch 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: Updating vm06:/etc/ceph/ceph.conf 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: Updating vm09:/etc/ceph/ceph.conf 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: Updating vm09:/var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/config/ceph.conf 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: Updating vm06:/var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/config/ceph.conf 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/1622960182' entity='client.admin' cmd=[{"prefix": "mon dump", "format": "json"}]: dispatch 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:47.207 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:46 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T10:01:47.262 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:47.545 INFO:teuthology.orchestra.run.vm06.stdout:# minimal ceph.conf for ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:01:47.545 INFO:teuthology.orchestra.run.vm06.stdout:[global] 2026-03-10T10:01:47.545 INFO:teuthology.orchestra.run.vm06.stdout: fsid = ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:01:47.545 INFO:teuthology.orchestra.run.vm06.stdout: mon_host = [v2:192.168.123.106:3300/0,v1:192.168.123.106:6789/0] [v2:192.168.123.109:3300/0,v1:192.168.123.109:6789/0] 2026-03-10T10:01:47.714 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-10T10:01:47.714 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T10:01:47.714 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T10:01:47.743 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T10:01:47.743 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:01:47.808 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:01:47.808 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.conf 2026-03-10T10:01:47.841 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:01:47.841 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-10T10:01:47.907 INFO:tasks.cephadm:Deploying OSDs... 2026-03-10T10:01:47.908 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T10:01:47.908 DEBUG:teuthology.orchestra.run.vm06:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T10:01:47.936 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:01:47.936 DEBUG:teuthology.orchestra.run.vm06:> ls /dev/[sv]d? 2026-03-10T10:01:47.995 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vda 2026-03-10T10:01:47.995 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdb 2026-03-10T10:01:47.995 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdc 2026-03-10T10:01:47.995 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vdd 2026-03-10T10:01:47.995 INFO:teuthology.orchestra.run.vm06.stdout:/dev/vde 2026-03-10T10:01:47.995 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T10:01:47.995 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T10:01:47.995 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdb 2026-03-10T10:01:48.054 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdb 2026-03-10T10:01:48.054 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:01:48.054 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 221 Links: 1 Device type: fc,10 2026-03-10T10:01:48.054 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:01:48.054 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:01:48.054 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 10:01:03.871776483 +0000 2026-03-10T10:01:48.054 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 09:59:44.473693115 +0000 2026-03-10T10:01:48.054 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 09:59:44.473693115 +0000 2026-03-10T10:01:48.054 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 09:56:25.271000000 +0000 2026-03-10T10:01:48.054 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T10:01:48.127 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T10:01:48.127 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T10:01:48.127 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 9.3444e-05 s, 5.5 MB/s 2026-03-10T10:01:48.128 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T10:01:48.193 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdc 2026-03-10T10:01:48.259 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdc 2026-03-10T10:01:48.259 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:01:48.259 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 224 Links: 1 Device type: fc,20 2026-03-10T10:01:48.259 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:01:48.259 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:01:48.259 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 10:01:03.923776569 +0000 2026-03-10T10:01:48.259 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 09:59:44.470693112 +0000 2026-03-10T10:01:48.259 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 09:59:44.470693112 +0000 2026-03-10T10:01:48.259 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 09:56:25.275000000 +0000 2026-03-10T10:01:48.259 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: Reconfiguring mon.vm06 (unknown last config time)... 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: Reconfiguring daemon mon.vm06 on vm06 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: Reconfiguring mgr.vm06.dkjjvn (unknown last config time)... 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm06.dkjjvn", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: Reconfiguring daemon mgr.vm06.dkjjvn on vm06 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1265314404' entity='client.admin' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm06", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm06", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T10:01:48.311 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:48 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:48.347 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T10:01:48.347 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T10:01:48.347 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000134242 s, 3.8 MB/s 2026-03-10T10:01:48.348 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T10:01:48.412 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vdd 2026-03-10T10:01:48.469 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vdd 2026-03-10T10:01:48.469 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:01:48.469 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 249 Links: 1 Device type: fc,30 2026-03-10T10:01:48.469 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:01:48.469 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:01:48.469 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 10:01:03.948776611 +0000 2026-03-10T10:01:48.469 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 09:59:44.461693103 +0000 2026-03-10T10:01:48.469 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 09:59:44.461693103 +0000 2026-03-10T10:01:48.469 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 09:56:25.290000000 +0000 2026-03-10T10:01:48.469 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T10:01:48.543 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T10:01:48.543 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T10:01:48.543 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000262031 s, 2.0 MB/s 2026-03-10T10:01:48.544 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T10:01:48.610 DEBUG:teuthology.orchestra.run.vm06:> stat /dev/vde 2026-03-10T10:01:48.667 INFO:teuthology.orchestra.run.vm06.stdout: File: /dev/vde 2026-03-10T10:01:48.667 INFO:teuthology.orchestra.run.vm06.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:01:48.667 INFO:teuthology.orchestra.run.vm06.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T10:01:48.667 INFO:teuthology.orchestra.run.vm06.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:01:48.667 INFO:teuthology.orchestra.run.vm06.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:01:48.667 INFO:teuthology.orchestra.run.vm06.stdout:Access: 2026-03-10 10:01:03.979776663 +0000 2026-03-10T10:01:48.667 INFO:teuthology.orchestra.run.vm06.stdout:Modify: 2026-03-10 09:59:44.472693114 +0000 2026-03-10T10:01:48.668 INFO:teuthology.orchestra.run.vm06.stdout:Change: 2026-03-10 09:59:44.472693114 +0000 2026-03-10T10:01:48.668 INFO:teuthology.orchestra.run.vm06.stdout: Birth: 2026-03-10 09:56:25.296000000 +0000 2026-03-10T10:01:48.668 DEBUG:teuthology.orchestra.run.vm06:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T10:01:48.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:48 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:48.744 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records in 2026-03-10T10:01:48.744 INFO:teuthology.orchestra.run.vm06.stderr:1+0 records out 2026-03-10T10:01:48.744 INFO:teuthology.orchestra.run.vm06.stderr:512 bytes copied, 0.000113733 s, 4.5 MB/s 2026-03-10T10:01:48.745 DEBUG:teuthology.orchestra.run.vm06:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T10:01:48.808 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:01:48.808 DEBUG:teuthology.orchestra.run.vm09:> dd if=/scratch_devs of=/dev/stdout 2026-03-10T10:01:48.825 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:01:48.825 DEBUG:teuthology.orchestra.run.vm09:> ls /dev/[sv]d? 2026-03-10T10:01:48.884 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vda 2026-03-10T10:01:48.884 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdb 2026-03-10T10:01:48.884 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdc 2026-03-10T10:01:48.884 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vdd 2026-03-10T10:01:48.884 INFO:teuthology.orchestra.run.vm09.stdout:/dev/vde 2026-03-10T10:01:48.884 WARNING:teuthology.misc:Removing root device: /dev/vda from device list 2026-03-10T10:01:48.884 DEBUG:teuthology.misc:devs=['/dev/vdb', '/dev/vdc', '/dev/vdd', '/dev/vde'] 2026-03-10T10:01:48.884 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdb 2026-03-10T10:01:48.943 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdb 2026-03-10T10:01:48.943 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:01:48.943 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 254 Links: 1 Device type: fc,10 2026-03-10T10:01:48.943 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:01:48.943 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:01:48.943 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 10:01:31.974837518 +0000 2026-03-10T10:01:48.943 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 09:59:43.156917977 +0000 2026-03-10T10:01:48.943 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 09:59:43.156917977 +0000 2026-03-10T10:01:48.943 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 09:55:54.291000000 +0000 2026-03-10T10:01:48.944 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdb of=/dev/null count=1 2026-03-10T10:01:49.009 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T10:01:49.009 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T10:01:49.009 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000190888 s, 2.7 MB/s 2026-03-10T10:01:49.010 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdb 2026-03-10T10:01:49.068 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdc 2026-03-10T10:01:49.127 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdc 2026-03-10T10:01:49.127 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:01:49.127 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 255 Links: 1 Device type: fc,20 2026-03-10T10:01:49.127 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:01:49.127 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:01:49.127 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 10:01:32.002837547 +0000 2026-03-10T10:01:49.127 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 09:59:43.164917987 +0000 2026-03-10T10:01:49.127 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 09:59:43.164917987 +0000 2026-03-10T10:01:49.127 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 09:55:54.298000000 +0000 2026-03-10T10:01:49.127 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdc of=/dev/null count=1 2026-03-10T10:01:49.194 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T10:01:49.195 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T10:01:49.195 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.0002351 s, 2.2 MB/s 2026-03-10T10:01:49.196 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdc 2026-03-10T10:01:49.257 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vdd 2026-03-10T10:01:49.318 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vdd 2026-03-10T10:01:49.319 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:01:49.319 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 256 Links: 1 Device type: fc,30 2026-03-10T10:01:49.319 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:01:49.319 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:01:49.319 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 10:01:32.029837574 +0000 2026-03-10T10:01:49.319 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 09:59:43.167917991 +0000 2026-03-10T10:01:49.319 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 09:59:43.167917991 +0000 2026-03-10T10:01:49.319 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 09:55:54.314000000 +0000 2026-03-10T10:01:49.319 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vdd of=/dev/null count=1 2026-03-10T10:01:49.384 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T10:01:49.384 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T10:01:49.384 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000198251 s, 2.6 MB/s 2026-03-10T10:01:49.385 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vdd 2026-03-10T10:01:49.445 DEBUG:teuthology.orchestra.run.vm09:> stat /dev/vde 2026-03-10T10:01:49.506 INFO:teuthology.orchestra.run.vm09.stdout: File: /dev/vde 2026-03-10T10:01:49.506 INFO:teuthology.orchestra.run.vm09.stdout: Size: 0 Blocks: 0 IO Block: 512 block special file 2026-03-10T10:01:49.506 INFO:teuthology.orchestra.run.vm09.stdout:Device: 6h/6d Inode: 257 Links: 1 Device type: fc,40 2026-03-10T10:01:49.506 INFO:teuthology.orchestra.run.vm09.stdout:Access: (0660/brw-rw----) Uid: ( 0/ root) Gid: ( 6/ disk) 2026-03-10T10:01:49.506 INFO:teuthology.orchestra.run.vm09.stdout:Context: system_u:object_r:fixed_disk_device_t:s0 2026-03-10T10:01:49.506 INFO:teuthology.orchestra.run.vm09.stdout:Access: 2026-03-10 10:01:32.058837604 +0000 2026-03-10T10:01:49.506 INFO:teuthology.orchestra.run.vm09.stdout:Modify: 2026-03-10 09:59:43.161917983 +0000 2026-03-10T10:01:49.506 INFO:teuthology.orchestra.run.vm09.stdout:Change: 2026-03-10 09:59:43.161917983 +0000 2026-03-10T10:01:49.506 INFO:teuthology.orchestra.run.vm09.stdout: Birth: 2026-03-10 09:55:54.370000000 +0000 2026-03-10T10:01:49.507 DEBUG:teuthology.orchestra.run.vm09:> sudo dd if=/dev/vde of=/dev/null count=1 2026-03-10T10:01:49.575 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records in 2026-03-10T10:01:49.575 INFO:teuthology.orchestra.run.vm09.stderr:1+0 records out 2026-03-10T10:01:49.575 INFO:teuthology.orchestra.run.vm09.stderr:512 bytes copied, 0.000213299 s, 2.4 MB/s 2026-03-10T10:01:49.576 DEBUG:teuthology.orchestra.run.vm09:> ! mount | grep -v devtmpfs | grep -q /dev/vde 2026-03-10T10:01:49.638 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch apply osd --all-available-devices 2026-03-10T10:01:49.704 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:49 vm09 ceph-mon[54558]: Reconfiguring ceph-exporter.vm06 (monmap changed)... 2026-03-10T10:01:49.704 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:49 vm09 ceph-mon[54558]: Reconfiguring daemon ceph-exporter.vm06 on vm06 2026-03-10T10:01:49.704 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:49 vm09 ceph-mon[54558]: Reconfiguring crash.vm06 (monmap changed)... 2026-03-10T10:01:49.704 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:49 vm09 ceph-mon[54558]: Reconfiguring daemon crash.vm06 on vm06 2026-03-10T10:01:49.704 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:49 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:49.704 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:49 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:49.704 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:49 vm09 ceph-mon[54558]: Reconfiguring alertmanager.vm06 (dependencies changed)... 2026-03-10T10:01:49.704 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:49 vm09 ceph-mon[54558]: Reconfiguring daemon alertmanager.vm06 on vm06 2026-03-10T10:01:49.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:49 vm06 ceph-mon[53281]: Reconfiguring ceph-exporter.vm06 (monmap changed)... 2026-03-10T10:01:49.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:49 vm06 ceph-mon[53281]: Reconfiguring daemon ceph-exporter.vm06 on vm06 2026-03-10T10:01:49.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:49 vm06 ceph-mon[53281]: Reconfiguring crash.vm06 (monmap changed)... 2026-03-10T10:01:49.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:49 vm06 ceph-mon[53281]: Reconfiguring daemon crash.vm06 on vm06 2026-03-10T10:01:49.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:49 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:49.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:49 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:49.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:49 vm06 ceph-mon[53281]: Reconfiguring alertmanager.vm06 (dependencies changed)... 2026-03-10T10:01:49.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:49 vm06 ceph-mon[53281]: Reconfiguring daemon alertmanager.vm06 on vm06 2026-03-10T10:01:49.870 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm09/config 2026-03-10T10:01:50.130 INFO:teuthology.orchestra.run.vm09.stdout:Scheduled osd.all-available-devices update... 2026-03-10T10:01:50.285 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-10T10:01:50.285 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:01:50.519 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:50.559 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:50 vm09 ceph-mon[54558]: pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:50.560 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:50 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:50.560 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:50 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:50.560 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:50 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:50.787 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:50.800 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:50 vm06 ceph-mon[53281]: pgmap v3: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:50.800 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:50 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:50.800 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:50 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:50.800 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:50 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:50.980 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-10T10:01:51.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:51 vm06 ceph-mon[53281]: Reconfiguring grafana.vm06 (dependencies changed)... 2026-03-10T10:01:51.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:51 vm06 ceph-mon[53281]: Reconfiguring daemon grafana.vm06 on vm06 2026-03-10T10:01:51.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:51 vm06 ceph-mon[53281]: from='client.14254 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:01:51.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:51 vm06 ceph-mon[53281]: Marking host: vm06 for OSDSpec preview refresh. 2026-03-10T10:01:51.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:51 vm06 ceph-mon[53281]: Marking host: vm09 for OSDSpec preview refresh. 2026-03-10T10:01:51.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:51 vm06 ceph-mon[53281]: Saving service osd.all-available-devices spec with placement * 2026-03-10T10:01:51.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:51 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3549450421' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:01:51.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:51 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:51.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:51 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:51.980 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:01:51.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:51 vm09 ceph-mon[54558]: Reconfiguring grafana.vm06 (dependencies changed)... 2026-03-10T10:01:51.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:51 vm09 ceph-mon[54558]: Reconfiguring daemon grafana.vm06 on vm06 2026-03-10T10:01:51.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:51 vm09 ceph-mon[54558]: from='client.14254 -' entity='client.admin' cmd=[{"prefix": "orch apply osd", "all_available_devices": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:01:51.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:51 vm09 ceph-mon[54558]: Marking host: vm06 for OSDSpec preview refresh. 2026-03-10T10:01:51.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:51 vm09 ceph-mon[54558]: Marking host: vm09 for OSDSpec preview refresh. 2026-03-10T10:01:51.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:51 vm09 ceph-mon[54558]: Saving service osd.all-available-devices spec with placement * 2026-03-10T10:01:51.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:51 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3549450421' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:01:51.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:51 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:51.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:51 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:52.287 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:52.553 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:52.713 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-10T10:01:52.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:52 vm06 ceph-mon[53281]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:52.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:52 vm06 ceph-mon[53281]: Reconfiguring prometheus.vm06 (dependencies changed)... 2026-03-10T10:01:52.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:52 vm06 ceph-mon[53281]: Reconfiguring daemon prometheus.vm06 on vm06 2026-03-10T10:01:52.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:52 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3620951362' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:01:52.832 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:52 vm09 ceph-mon[54558]: pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:52.832 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:52 vm09 ceph-mon[54558]: Reconfiguring prometheus.vm06 (dependencies changed)... 2026-03-10T10:01:52.833 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:52 vm09 ceph-mon[54558]: Reconfiguring daemon prometheus.vm06 on vm06 2026-03-10T10:01:52.833 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:52 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3620951362' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:01:53.713 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:01:53.893 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: Reconfiguring ceph-exporter.vm09 (monmap changed)... 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: Reconfiguring daemon ceph-exporter.vm09 on vm09 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.jzyfcu", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:01:53.916 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:53 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:53.933 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.933 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.933 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: Reconfiguring ceph-exporter.vm09 (monmap changed)... 2026-03-10T10:01:53.933 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.ceph-exporter.vm09", "caps": ["mon", "profile ceph-exporter", "mon", "allow r", "mgr", "allow r", "osd", "allow r"]}]: dispatch 2026-03-10T10:01:53.933 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:53.933 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: Reconfiguring daemon ceph-exporter.vm09 on vm09 2026-03-10T10:01:53.933 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.933 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.933 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.crash.vm09", "caps": ["mon", "profile crash", "mgr", "profile crash"]}]: dispatch 2026-03-10T10:01:53.934 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:53.934 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.934 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:53.934 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "mgr.vm09.jzyfcu", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]}]: dispatch 2026-03-10T10:01:53.934 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mgr services"}]: dispatch 2026-03-10T10:01:53.934 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:53 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:54.132 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:54.702 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: Reconfiguring crash.vm09 (monmap changed)... 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: Reconfiguring daemon crash.vm09 on vm09 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: Reconfiguring mgr.vm09.jzyfcu (monmap changed)... 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: Reconfiguring daemon mgr.vm09.jzyfcu on vm09 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/238907891' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm06.local:9093"}]: dispatch 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm06.local:3000"}]: dispatch 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm06.local:9095"}]: dispatch 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:54 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: Reconfiguring crash.vm09 (monmap changed)... 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: Reconfiguring daemon crash.vm09 on vm09 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: Reconfiguring mgr.vm09.jzyfcu (monmap changed)... 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: Reconfiguring daemon mgr.vm09.jzyfcu on vm09 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "mon."}]: dispatch 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config get", "who": "mon", "key": "public_network"}]: dispatch 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/238907891' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm06.local:9093"}]: dispatch 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm06.local:3000"}]: dispatch 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm06.local:9095"}]: dispatch 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:54.848 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:54 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:01:55.703 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: Reconfiguring mon.vm09 (monmap changed)... 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: Reconfiguring daemon mon.vm09 on vm09 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm06.local:9093"}]: dispatch 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm06.local:3000"}]: dispatch 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm06.local:9095"}]: dispatch 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:55.739 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:55 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:55.948 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: Reconfiguring mon.vm09 (monmap changed)... 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: Reconfiguring daemon mon.vm09 on vm09 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-alertmanager-api-host"}]: dispatch 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-alertmanager-api-host", "value": "http://vm06.local:9093"}]: dispatch 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-grafana-api-url"}]: dispatch 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-grafana-api-url", "value": "https://vm06.local:3000"}]: dispatch 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-prometheus-api-host", "value": "http://vm06.local:9095"}]: dispatch 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:55.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:55 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:56.245 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:56.421 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":5,"num_osds":0,"num_up_osds":0,"osd_up_since":0,"num_in_osds":0,"osd_in_since":0,"num_remapped_pgs":0} 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:57.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:56 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/474611691' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.bootstrap-osd"}]: dispatch 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:01:57.172 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:56 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/474611691' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:01:57.421 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:01:57.631 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:57.905 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: from='client.? 192.168.123.109:0/1017167265' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c6441cba-6366-4f4b-a30e-5f30a7196eb8"}]: dispatch 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c6441cba-6366-4f4b-a30e-5f30a7196eb8"}]: dispatch 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c6441cba-6366-4f4b-a30e-5f30a7196eb8"}]': finished 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3736492330' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "984ffa57-5df4-46a4-acfb-9d25f1ce2c76"}]: dispatch 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3736492330' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "984ffa57-5df4-46a4-acfb-9d25f1ce2c76"}]': finished 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: osdmap e7: 2 total, 0 up, 2 in 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:01:57.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:57 vm09 ceph-mon[54558]: from='client.? 192.168.123.109:0/1926625006' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:01:58.032 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/1017167265' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c6441cba-6366-4f4b-a30e-5f30a7196eb8"}]: dispatch 2026-03-10T10:01:58.032 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c6441cba-6366-4f4b-a30e-5f30a7196eb8"}]: dispatch 2026-03-10T10:01:58.032 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c6441cba-6366-4f4b-a30e-5f30a7196eb8"}]': finished 2026-03-10T10:01:58.032 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:01:58.032 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: osdmap e6: 1 total, 0 up, 1 in 2026-03-10T10:01:58.032 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:01:58.032 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3736492330' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "984ffa57-5df4-46a4-acfb-9d25f1ce2c76"}]: dispatch 2026-03-10T10:01:58.032 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3736492330' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "984ffa57-5df4-46a4-acfb-9d25f1ce2c76"}]': finished 2026-03-10T10:01:58.032 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: osdmap e7: 2 total, 0 up, 2 in 2026-03-10T10:01:58.032 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:01:58.032 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:01:58.033 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:57 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/1926625006' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:01:58.064 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":7,"num_osds":2,"num_up_osds":0,"osd_up_since":0,"num_in_osds":2,"osd_in_since":1773136917,"num_remapped_pgs":0} 2026-03-10T10:01:59.065 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:01:59.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:58 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/274365678' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:01:59.067 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:58 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1891932905' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:01:59.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:58 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/274365678' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:01:59.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:58 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1891932905' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:01:59.247 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:01:59.508 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:01:59.691 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":7,"num_osds":2,"num_up_osds":0,"osd_up_since":0,"num_in_osds":2,"osd_in_since":1773136917,"num_remapped_pgs":0} 2026-03-10T10:02:00.024 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:59 vm06 ceph-mon[53281]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:00.024 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:59 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:02:00.024 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:01:59 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3806948029' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:00.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:59 vm09 ceph-mon[54558]: pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:00.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:59 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:02:00.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:01:59 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3806948029' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:00.692 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:00.903 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:00.945 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:00 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/757319084' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b830bc44-e908-40b0-aa15-2527e834436e"}]: dispatch 2026-03-10T10:02:00.945 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:00 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/757319084' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b830bc44-e908-40b0-aa15-2527e834436e"}]': finished 2026-03-10T10:02:00.945 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:00 vm09 ceph-mon[54558]: osdmap e8: 3 total, 0 up, 3 in 2026-03-10T10:02:00.945 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:00 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:00.945 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:00 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:00.945 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:00 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:01.080 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:00 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/757319084' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "b830bc44-e908-40b0-aa15-2527e834436e"}]: dispatch 2026-03-10T10:02:01.080 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:00 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/757319084' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "b830bc44-e908-40b0-aa15-2527e834436e"}]': finished 2026-03-10T10:02:01.080 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:00 vm06 ceph-mon[53281]: osdmap e8: 3 total, 0 up, 3 in 2026-03-10T10:02:01.080 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:00 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:01.080 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:00 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:01.080 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:00 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:01.195 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:01.348 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":9,"num_osds":4,"num_up_osds":0,"osd_up_since":0,"num_in_osds":4,"osd_in_since":1773136920,"num_remapped_pgs":0} 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: from='client.? 192.168.123.109:0/462933155' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "51747f43-8d7c-498e-bab3-fd72ea514f53"}]: dispatch 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "51747f43-8d7c-498e-bab3-fd72ea514f53"}]: dispatch 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "51747f43-8d7c-498e-bab3-fd72ea514f53"}]': finished 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: osdmap e9: 4 total, 0 up, 4 in 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/533117927' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1875932346' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:01.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:01 vm09 ceph-mon[54558]: from='client.? 192.168.123.109:0/355024737' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/462933155' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "51747f43-8d7c-498e-bab3-fd72ea514f53"}]: dispatch 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "51747f43-8d7c-498e-bab3-fd72ea514f53"}]: dispatch 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "51747f43-8d7c-498e-bab3-fd72ea514f53"}]': finished 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: osdmap e9: 4 total, 0 up, 4 in 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/533117927' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1875932346' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:02.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:01 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/355024737' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:02.349 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:02.546 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:02.806 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:02.901 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:02 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3843509134' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:02.986 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":9,"num_osds":4,"num_up_osds":0,"osd_up_since":0,"num_in_osds":4,"osd_in_since":1773136920,"num_remapped_pgs":0} 2026-03-10T10:02:03.239 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:02 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3843509134' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:03.988 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:04.164 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:04.192 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:03 vm06 ceph-mon[53281]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:04.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:03 vm09 ceph-mon[54558]: pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:04.400 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:04.549 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":9,"num_osds":4,"num_up_osds":0,"osd_up_since":0,"num_in_osds":4,"osd_in_since":1773136920,"num_remapped_pgs":0} 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3678996619' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/573295685' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3937fe45-31f7-4447-9ada-32834c16699a"}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/573295685' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "3937fe45-31f7-4447-9ada-32834c16699a"}]': finished 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: osdmap e10: 5 total, 0 up, 5 in 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/4110743702' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e0feac48-24ef-487a-af9a-a863a28ad6e1"}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e0feac48-24ef-487a-af9a-a863a28ad6e1"}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e0feac48-24ef-487a-af9a-a863a28ad6e1"}]': finished 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: osdmap e11: 6 total, 0 up, 6 in 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:05.229 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:04 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:05.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3678996619' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/573295685' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "3937fe45-31f7-4447-9ada-32834c16699a"}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/573295685' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "3937fe45-31f7-4447-9ada-32834c16699a"}]': finished 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: osdmap e10: 5 total, 0 up, 5 in 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='client.? 192.168.123.109:0/4110743702' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e0feac48-24ef-487a-af9a-a863a28ad6e1"}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "e0feac48-24ef-487a-af9a-a863a28ad6e1"}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "e0feac48-24ef-487a-af9a-a863a28ad6e1"}]': finished 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: osdmap e11: 6 total, 0 up, 6 in 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:05.241 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:04 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:05.550 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:05.716 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:05.956 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:06.130 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:05 vm06 ceph-mon[53281]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:06.130 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:05 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/382720611' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:06.130 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:05 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/2823320119' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:06.157 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":11,"num_osds":6,"num_up_osds":0,"osd_up_since":0,"num_in_osds":6,"osd_in_since":1773136924,"num_remapped_pgs":0} 2026-03-10T10:02:06.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:05 vm09 ceph-mon[54558]: pgmap v17: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:06.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:05 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/382720611' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:06.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:05 vm09 ceph-mon[54558]: from='client.? 192.168.123.109:0/2823320119' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:07.158 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:07.180 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:06 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3681762679' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:07.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:06 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3681762679' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:07.326 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:07.554 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:07.708 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":11,"num_osds":6,"num_up_osds":0,"osd_up_since":0,"num_in_osds":6,"osd_in_since":1773136924,"num_remapped_pgs":0} 2026-03-10T10:02:07.987 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:07 vm06 ceph-mon[53281]: pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:07.987 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:07 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1921038324' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:08.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:07 vm09 ceph-mon[54558]: pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:08.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:07 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1921038324' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:08.708 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:08.890 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:09.149 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1099327753' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c9a065b3-6872-43ce-a26f-ba45448813d8"}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1099327753' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c9a065b3-6872-43ce-a26f-ba45448813d8"}]': finished 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: osdmap e12: 7 total, 0 up, 7 in 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/2350390168' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a1f6b136-68c5-4bc2-909e-86f730593029"}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a1f6b136-68c5-4bc2-909e-86f730593029"}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a1f6b136-68c5-4bc2-909e-86f730593029"}]': finished 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: osdmap e13: 8 total, 0 up, 8 in 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:09.182 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:09.183 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:09.183 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:09.183 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:09.183 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:08 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/565933338' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1099327753' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "c9a065b3-6872-43ce-a26f-ba45448813d8"}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1099327753' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c9a065b3-6872-43ce-a26f-ba45448813d8"}]': finished 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: osdmap e12: 7 total, 0 up, 7 in 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='client.? 192.168.123.109:0/2350390168' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a1f6b136-68c5-4bc2-909e-86f730593029"}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='client.? ' entity='client.bootstrap-osd' cmd=[{"prefix": "osd new", "uuid": "a1f6b136-68c5-4bc2-909e-86f730593029"}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a1f6b136-68c5-4bc2-909e-86f730593029"}]': finished 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: osdmap e13: 8 total, 0 up, 8 in 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:09.227 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:08 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/565933338' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:09.322 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":0} 2026-03-10T10:02:10.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:09 vm09 ceph-mon[54558]: pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:10.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:09 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1129047877' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:10.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:09 vm09 ceph-mon[54558]: from='client.? 192.168.123.109:0/542243910' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:10.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:09 vm06 ceph-mon[53281]: pgmap v21: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:10.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:09 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1129047877' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:10.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:09 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/542243910' entity='client.bootstrap-osd' cmd=[{"prefix": "mon getmap"}]: dispatch 2026-03-10T10:02:10.322 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:10.498 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:10.745 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:10.893 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":0} 2026-03-10T10:02:11.118 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:10 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/451947537' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:11.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:10 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/451947537' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:11.894 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:12.049 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:11 vm06 ceph-mon[53281]: pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:12.112 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:12.236 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:11 vm09 ceph-mon[54558]: pgmap v22: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:12.361 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:12.548 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":0} 2026-03-10T10:02:12.951 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:12 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1551452873' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:12.951 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:12 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:02:12.951 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:12 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:12.951 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:12 vm06 ceph-mon[53281]: Deploying daemon osd.1 on vm06 2026-03-10T10:02:13.234 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:12 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1551452873' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:13.234 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:12 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.1"}]: dispatch 2026-03-10T10:02:13.234 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:12 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:13.234 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:12 vm09 ceph-mon[54558]: Deploying daemon osd.1 on vm06 2026-03-10T10:02:13.549 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:13.883 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:14.023 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:13 vm09 ceph-mon[54558]: pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:14.023 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:13 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:02:14.023 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:13 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:14.023 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:13 vm09 ceph-mon[54558]: Deploying daemon osd.0 on vm09 2026-03-10T10:02:14.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:13 vm06 ceph-mon[53281]: pgmap v23: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:14.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:13 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.0"}]: dispatch 2026-03-10T10:02:14.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:13 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:14.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:13 vm06 ceph-mon[53281]: Deploying daemon osd.0 on vm09 2026-03-10T10:02:14.265 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:14.463 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":0} 2026-03-10T10:02:15.126 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:14 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:02:15.126 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:14 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/2497509774' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:15.430 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:14 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:02:15.430 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:14 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/2497509774' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:15.464 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:15.703 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:15.972 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:16.131 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":13,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":0} 2026-03-10T10:02:16.267 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:16 vm06 ceph-mon[53281]: pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:16.267 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:16 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:16.267 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:16 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:16.267 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:16 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:02:16.267 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:16 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:16.267 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:16 vm06 ceph-mon[53281]: Deploying daemon osd.2 on vm06 2026-03-10T10:02:16.267 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:16 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:16.267 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:16 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:16.267 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:16 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T10:02:16.267 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:16 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:16.267 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:16 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/234659950' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:16.287 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:16 vm09 ceph-mon[54558]: pgmap v24: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:16.287 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:16 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:16.287 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:16 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:16.287 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:16 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.2"}]: dispatch 2026-03-10T10:02:16.287 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:16 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:16.287 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:16 vm09 ceph-mon[54558]: Deploying daemon osd.2 on vm06 2026-03-10T10:02:16.287 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:16 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:16.287 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:16 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:16.287 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:16 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.3"}]: dispatch 2026-03-10T10:02:16.287 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:16 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:16.287 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:16 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/234659950' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:17.131 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:17.355 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:17.391 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:17 vm09 ceph-mon[54558]: Deploying daemon osd.3 on vm09 2026-03-10T10:02:17.391 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:17 vm09 ceph-mon[54558]: from='osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:02:17.413 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:17 vm06 ceph-mon[53281]: Deploying daemon osd.3 on vm09 2026-03-10T10:02:17.413 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:17 vm06 ceph-mon[53281]: from='osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501]' entity='osd.1' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]: dispatch 2026-03-10T10:02:17.623 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:17.787 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":14,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":0} 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: osdmap e14: 8 total, 0 up, 8 in 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='osd.0 [v2:192.168.123.109:6800/1572520565,v1:192.168.123.109:6801/1572520565]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:02:18.267 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3215077434' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:18.268 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:18.268 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:18.268 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T10:02:18.268 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:18 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: pgmap v25: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: osdmap e14: 8 total, 0 up, 8 in 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501]' entity='osd.1' cmd=[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='osd.0 [v2:192.168.123.109:6800/1572520565,v1:192.168.123.109:6801/1572520565]' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3215077434' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.4"}]: dispatch 2026-03-10T10:02:18.521 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:18 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:18.788 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:19.092 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:19.413 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: Deploying daemon osd.4 on vm06 2026-03-10T10:02:19.413 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T10:02:19.413 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T10:02:19.413 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: osdmap e15: 8 total, 0 up, 8 in 2026-03-10T10:02:19.413 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='osd.0 [v2:192.168.123.109:6800/1572520565,v1:192.168.123.109:6801/1572520565]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:19.413 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:19.413 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: Deploying daemon osd.5 on vm09 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:02:19.414 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:19 vm06 ceph-mon[53281]: osdmap e16: 8 total, 0 up, 8 in 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: Deploying daemon osd.4 on vm06 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501]' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: osdmap e15: 8 total, 0 up, 8 in 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='osd.0 [v2:192.168.123.109:6800/1572520565,v1:192.168.123.109:6801/1572520565]' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='osd.0 ' entity='osd.0' cmd=[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.5"}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: Deploying daemon osd.5 on vm09 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: from='osd.0 ' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:02:19.494 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:19 vm09 ceph-mon[54558]: osdmap e16: 8 total, 0 up, 8 in 2026-03-10T10:02:19.593 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:19.755 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":16,"num_osds":8,"num_up_osds":0,"osd_up_since":0,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":0} 2026-03-10T10:02:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: purged_snaps scrub starts 2026-03-10T10:02:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: purged_snaps scrub ok 2026-03-10T10:02:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:20.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:20.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:20.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501]' entity='osd.1' 2026-03-10T10:02:20.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/4180725465' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:20.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='osd.2 [v2:192.168.123.106:6810/4167697000,v1:192.168.123.106:6811/4167697000]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:02:20.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='osd.3 [v2:192.168.123.109:6808/3783651379,v1:192.168.123.109:6809/3783651379]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T10:02:20.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T10:02:20.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:20.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:20 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:20.698 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: purged_snaps scrub starts 2026-03-10T10:02:20.698 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: purged_snaps scrub ok 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: pgmap v28: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501]' entity='osd.1' 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/4180725465' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='osd.2 [v2:192.168.123.106:6810/4167697000,v1:192.168.123.106:6811/4167697000]' entity='osd.2' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='osd.3 [v2:192.168.123.109:6808/3783651379,v1:192.168.123.109:6809/3783651379]' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:20.699 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:20 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:20.756 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:20.964 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:21.316 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: purged_snaps scrub starts 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: purged_snaps scrub ok 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='osd.2 [v2:192.168.123.106:6810/4167697000,v1:192.168.123.106:6811/4167697000]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: osd.0 [v2:192.168.123.109:6800/1572520565,v1:192.168.123.109:6801/1572520565] boot 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501] boot 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: osdmap e17: 8 total, 2 up, 8 in 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='osd.2 [v2:192.168.123.106:6810/4167697000,v1:192.168.123.106:6811/4167697000]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='osd.3 [v2:192.168.123.109:6808/3783651379,v1:192.168.123.109:6809/3783651379]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='osd.2 [v2:192.168.123.106:6810/4167697000,v1:192.168.123.106:6811/4167697000]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:02:21.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:21 vm06 ceph-mon[53281]: osdmap e18: 8 total, 2 up, 8 in 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: purged_snaps scrub starts 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: purged_snaps scrub ok 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='osd.2 [v2:192.168.123.106:6810/4167697000,v1:192.168.123.106:6811/4167697000]' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: osd.0 [v2:192.168.123.109:6800/1572520565,v1:192.168.123.109:6801/1572520565] boot 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: osd.1 [v2:192.168.123.106:6802/1902557501,v1:192.168.123.106:6803/1902557501] boot 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: osdmap e17: 8 total, 2 up, 8 in 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='osd.2 [v2:192.168.123.106:6810/4167697000,v1:192.168.123.106:6811/4167697000]' entity='osd.2' cmd=[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 0}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 1}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='osd.3 [v2:192.168.123.109:6808/3783651379,v1:192.168.123.109:6809/3783651379]' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='osd.3 ' entity='osd.3' cmd=[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.6"}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "osd.7"}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='osd.2 [v2:192.168.123.106:6810/4167697000,v1:192.168.123.106:6811/4167697000]' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:02:21.574 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:21 vm09 ceph-mon[54558]: osdmap e18: 8 total, 2 up, 8 in 2026-03-10T10:02:21.599 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":18,"num_osds":8,"num_up_osds":2,"osd_up_since":1773136940,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":0} 2026-03-10T10:02:22.453 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: Deploying daemon osd.6 on vm06 2026-03-10T10:02:22.453 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: Deploying daemon osd.7 on vm09 2026-03-10T10:02:22.454 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:22.454 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:22.454 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:22.454 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:22.454 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:22.454 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:22.454 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:22.454 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/2048483900' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:22.454 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:22.454 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:22 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:22.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: Deploying daemon osd.6 on vm06 2026-03-10T10:02:22.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: Deploying daemon osd.7 on vm09 2026-03-10T10:02:22.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: pgmap v31: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-10T10:02:22.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:22.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:22.565 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:22.566 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:22.566 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:22.566 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:22.566 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/2048483900' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:22.566 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:22.566 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:22 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:22.600 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:22.883 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:23.195 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:23.437 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":19,"num_osds":8,"num_up_osds":3,"osd_up_since":1773136942,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":0} 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: purged_snaps scrub starts 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: purged_snaps scrub ok 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: purged_snaps scrub starts 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: purged_snaps scrub ok 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: osd.2 [v2:192.168.123.106:6810/4167697000,v1:192.168.123.106:6811/4167697000] boot 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: osdmap e19: 8 total, 3 up, 8 in 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='osd.4 [v2:192.168.123.106:6818/1691146704,v1:192.168.123.106:6819/1691146704]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='osd.3 ' entity='osd.3' 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='osd.5 [v2:192.168.123.109:6816/87938156,v1:192.168.123.109:6817/87938156]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/4005701920' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:23.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:23 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:23.602 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: purged_snaps scrub starts 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: purged_snaps scrub ok 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: purged_snaps scrub starts 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: purged_snaps scrub ok 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: osd.2 [v2:192.168.123.106:6810/4167697000,v1:192.168.123.106:6811/4167697000] boot 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: osdmap e19: 8 total, 3 up, 8 in 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 2}]: dispatch 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='osd.4 [v2:192.168.123.106:6818/1691146704,v1:192.168.123.106:6819/1691146704]' entity='osd.4' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]: dispatch 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='osd.3 ' entity='osd.3' 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='osd.5 [v2:192.168.123.109:6816/87938156,v1:192.168.123.109:6817/87938156]' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]: dispatch 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/4005701920' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:23.603 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:23 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:24.439 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:24.442 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:02:24.442 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='osd.4 [v2:192.168.123.106:6818/1691146704,v1:192.168.123.106:6819/1691146704]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T10:02:24.442 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T10:02:24.442 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T10:02:24.442 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: osd.3 [v2:192.168.123.109:6808/3783651379,v1:192.168.123.109:6809/3783651379] boot 2026-03-10T10:02:24.442 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: osdmap e20: 8 total, 4 up, 8 in 2026-03-10T10:02:24.442 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='osd.4 [v2:192.168.123.106:6818/1691146704,v1:192.168.123.106:6819/1691146704]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='osd.5 [v2:192.168.123.109:6816/87938156,v1:192.168.123.109:6817/87938156]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:24.443 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:24 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:24.467 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: pgmap v34: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='osd.4 [v2:192.168.123.106:6818/1691146704,v1:192.168.123.106:6819/1691146704]' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: osd.3 [v2:192.168.123.109:6808/3783651379,v1:192.168.123.109:6809/3783651379] boot 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: osdmap e20: 8 total, 4 up, 8 in 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='osd.4 [v2:192.168.123.106:6818/1691146704,v1:192.168.123.106:6819/1691146704]' entity='osd.4' cmd=[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 3}]: dispatch 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='osd.5 [v2:192.168.123.109:6816/87938156,v1:192.168.123.109:6817/87938156]' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='osd.5 ' entity='osd.5' cmd=[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:24.468 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:24 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:24.732 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:25.107 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:25.332 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":21,"num_osds":8,"num_up_osds":4,"osd_up_since":1773136943,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":0} 2026-03-10T10:02:25.383 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:25 vm09 ceph-mon[54558]: from='osd.4 [v2:192.168.123.106:6818/1691146704,v1:192.168.123.106:6819/1691146704]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T10:02:25.383 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:25 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:02:25.383 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:25 vm09 ceph-mon[54558]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:02:25.383 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:25 vm09 ceph-mon[54558]: osdmap e21: 8 total, 4 up, 8 in 2026-03-10T10:02:25.383 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:25 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:25.383 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:25 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:25.383 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:25 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:25.383 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:25 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:25.383 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:25 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:25.383 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:25 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3345377518' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:25.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:25 vm06 ceph-mon[53281]: from='osd.4 [v2:192.168.123.106:6818/1691146704,v1:192.168.123.106:6819/1691146704]' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T10:02:25.569 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:25 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:02:25.570 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:25 vm06 ceph-mon[53281]: from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:02:25.570 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:25 vm06 ceph-mon[53281]: osdmap e21: 8 total, 4 up, 8 in 2026-03-10T10:02:25.570 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:25 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:25.570 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:25 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:25.570 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:25 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:25.570 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:25 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:25.570 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:25 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:25.570 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:25 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3345377518' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:26.334 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: purged_snaps scrub starts 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: purged_snaps scrub ok 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: purged_snaps scrub starts 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: purged_snaps scrub ok 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: pgmap v37: 1 pgs: 1 creating+peering; 0 B data, 105 MiB used, 80 GiB / 80 GiB avail 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: osd.4 [v2:192.168.123.106:6818/1691146704,v1:192.168.123.106:6819/1691146704] boot 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: osdmap e22: 8 total, 5 up, 8 in 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='osd.7 [v2:192.168.123.109:6824/2024432096,v1:192.168.123.109:6825/2024432096]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='osd.5 ' entity='osd.5' 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='osd.6 [v2:192.168.123.106:6826/2533846059,v1:192.168.123.106:6827/2533846059]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='osd.6 [v2:192.168.123.106:6826/2533846059,v1:192.168.123.106:6827/2533846059]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: osd.5 [v2:192.168.123.109:6816/87938156,v1:192.168.123.109:6817/87938156] boot 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: osdmap e23: 8 total, 6 up, 8 in 2026-03-10T10:02:26.371 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='osd.6 [v2:192.168.123.106:6826/2533846059,v1:192.168.123.106:6827/2533846059]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T10:02:26.372 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:26.372 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:26.372 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:26.372 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='osd.7 [v2:192.168.123.109:6824/2024432096,v1:192.168.123.109:6825/2024432096]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:26.372 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:26.372 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.372 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:26 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.581 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: purged_snaps scrub starts 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: purged_snaps scrub ok 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: purged_snaps scrub starts 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: purged_snaps scrub ok 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: pgmap v37: 1 pgs: 1 creating+peering; 0 B data, 105 MiB used, 80 GiB / 80 GiB avail 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: osd.4 [v2:192.168.123.106:6818/1691146704,v1:192.168.123.106:6819/1691146704] boot 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: osdmap e22: 8 total, 5 up, 8 in 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 4}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='osd.7 [v2:192.168.123.109:6824/2024432096,v1:192.168.123.109:6825/2024432096]' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='osd.5 ' entity='osd.5' 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='osd.6 [v2:192.168.123.106:6826/2533846059,v1:192.168.123.106:6827/2533846059]' entity='osd.6' cmd=[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='osd.6 [v2:192.168.123.106:6826/2533846059,v1:192.168.123.106:6827/2533846059]' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: osd.5 [v2:192.168.123.109:6816/87938156,v1:192.168.123.109:6817/87938156] boot 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: osdmap e23: 8 total, 6 up, 8 in 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='osd.6 [v2:192.168.123.106:6826/2533846059,v1:192.168.123.106:6827/2533846059]' entity='osd.6' cmd=[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm06", "root=default"]}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 5}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='osd.7 [v2:192.168.123.109:6824/2024432096,v1:192.168.123.109:6825/2024432096]' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='osd.7 ' entity='osd.7' cmd=[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]: dispatch 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.659 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:26 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:26.895 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:27.094 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":24,"num_osds":8,"num_up_osds":6,"osd_up_since":1773136945,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":0} 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='osd.6 [v2:192.168.123.106:6826/2533846059,v1:192.168.123.106:6827/2533846059]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: osdmap e24: 8 total, 6 up, 8 in 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3125911582' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: pgmap v41: 1 pgs: 1 creating+peering; 0 B data, 158 MiB used, 120 GiB / 120 GiB avail 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: Detected new or changed devices on vm09 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:02:27.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:27 vm09 ceph-mon[54558]: from='osd.7 ' entity='osd.7' 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='osd.6 [v2:192.168.123.106:6826/2533846059,v1:192.168.123.106:6827/2533846059]' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm06", "root=default"]}]': finished 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm09", "root=default"]}]': finished 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: osdmap e24: 8 total, 6 up, 8 in 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3125911582' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: pgmap v41: 1 pgs: 1 creating+peering; 0 B data, 158 MiB used, 120 GiB / 120 GiB avail 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: Detected new or changed devices on vm09 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config rm", "who": "osd/host:vm09", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:02:28.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:27 vm06 ceph-mon[53281]: from='osd.7 ' entity='osd.7' 2026-03-10T10:02:28.095 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd stat -f json 2026-03-10T10:02:28.323 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:28.583 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:28.753 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":25,"num_osds":8,"num_up_osds":8,"osd_up_since":1773136947,"num_in_osds":8,"osd_in_since":1773136928,"num_remapped_pgs":1} 2026-03-10T10:02:28.753 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd dump --format=json 2026-03-10T10:02:28.938 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:28.964 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: purged_snaps scrub starts 2026-03-10T10:02:28.964 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: purged_snaps scrub ok 2026-03-10T10:02:28.964 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: purged_snaps scrub starts 2026-03-10T10:02:28.964 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: purged_snaps scrub ok 2026-03-10T10:02:28.964 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:28.964 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:28.964 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: osd.6 [v2:192.168.123.106:6826/2533846059,v1:192.168.123.106:6827/2533846059] boot 2026-03-10T10:02:28.964 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: osd.7 [v2:192.168.123.109:6824/2024432096,v1:192.168.123.109:6825/2024432096] boot 2026-03-10T10:02:28.964 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: osdmap e25: 8 total, 8 up, 8 in 2026-03-10T10:02:28.964 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:28.964 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:28.965 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: Detected new or changed devices on vm06 2026-03-10T10:02:28.965 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:28.965 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:28.965 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:02:28.965 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:28.965 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:02:28.965 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:28.965 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:02:28.965 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:28 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/188987565' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: purged_snaps scrub starts 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: purged_snaps scrub ok 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: purged_snaps scrub starts 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: purged_snaps scrub ok 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: osd.6 [v2:192.168.123.106:6826/2533846059,v1:192.168.123.106:6827/2533846059] boot 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: osd.7 [v2:192.168.123.109:6824/2024432096,v1:192.168.123.109:6825/2024432096] boot 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: osdmap e25: 8 total, 8 up, 8 in 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 6}]: dispatch 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd metadata", "id": 7}]: dispatch 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: Detected new or changed devices on vm06 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config rm", "who": "osd/host:vm06", "name": "osd_memory_target"}]: dispatch 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:02:28.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:28 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/188987565' entity='client.admin' cmd=[{"prefix": "osd stat", "format": "json"}]: dispatch 2026-03-10T10:02:29.197 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:29.197 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":26,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","created":"2026-03-10T10:00:34.057913+0000","modified":"2026-03-10T10:02:28.727602+0000","last_up_change":"2026-03-10T10:02:27.714365+0000","last_in_change":"2026-03-10T10:02:08.475060+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":11,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T10:02:23.152169+0000","flags":32769,"flags_names":"hashpspool,creating","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"21","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"c6441cba-6366-4f4b-a30e-5f30a7196eb8","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6801","nonce":1572520565}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6803","nonce":1572520565}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6807","nonce":1572520565}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6805","nonce":1572520565}]},"public_addr":"192.168.123.109:6801/1572520565","cluster_addr":"192.168.123.109:6803/1572520565","heartbeat_back_addr":"192.168.123.109:6807/1572520565","heartbeat_front_addr":"192.168.123.109:6805/1572520565","state":["exists","up"]},{"osd":1,"uuid":"984ffa57-5df4-46a4-acfb-9d25f1ce2c76","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6803","nonce":1902557501}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6805","nonce":1902557501}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6809","nonce":1902557501}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6807","nonce":1902557501}]},"public_addr":"192.168.123.106:6803/1902557501","cluster_addr":"192.168.123.106:6805/1902557501","heartbeat_back_addr":"192.168.123.106:6809/1902557501","heartbeat_front_addr":"192.168.123.106:6807/1902557501","state":["exists","up"]},{"osd":2,"uuid":"b830bc44-e908-40b0-aa15-2527e834436e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6811","nonce":4167697000}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6813","nonce":4167697000}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6817","nonce":4167697000}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6815","nonce":4167697000}]},"public_addr":"192.168.123.106:6811/4167697000","cluster_addr":"192.168.123.106:6813/4167697000","heartbeat_back_addr":"192.168.123.106:6817/4167697000","heartbeat_front_addr":"192.168.123.106:6815/4167697000","state":["exists","up"]},{"osd":3,"uuid":"51747f43-8d7c-498e-bab3-fd72ea514f53","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":23,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6809","nonce":3783651379}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6811","nonce":3783651379}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6815","nonce":3783651379}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6813","nonce":3783651379}]},"public_addr":"192.168.123.109:6809/3783651379","cluster_addr":"192.168.123.109:6811/3783651379","heartbeat_back_addr":"192.168.123.109:6815/3783651379","heartbeat_front_addr":"192.168.123.109:6813/3783651379","state":["exists","up"]},{"osd":4,"uuid":"3937fe45-31f7-4447-9ada-32834c16699a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":22,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6819","nonce":1691146704}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6821","nonce":1691146704}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6825","nonce":1691146704}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6823","nonce":1691146704}]},"public_addr":"192.168.123.106:6819/1691146704","cluster_addr":"192.168.123.106:6821/1691146704","heartbeat_back_addr":"192.168.123.106:6825/1691146704","heartbeat_front_addr":"192.168.123.106:6823/1691146704","state":["exists","up"]},{"osd":5,"uuid":"e0feac48-24ef-487a-af9a-a863a28ad6e1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6817","nonce":87938156}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6819","nonce":87938156}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6823","nonce":87938156}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6821","nonce":87938156}]},"public_addr":"192.168.123.109:6817/87938156","cluster_addr":"192.168.123.109:6819/87938156","heartbeat_back_addr":"192.168.123.109:6823/87938156","heartbeat_front_addr":"192.168.123.109:6821/87938156","state":["exists","up"]},{"osd":6,"uuid":"c9a065b3-6872-43ce-a26f-ba45448813d8","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6827","nonce":2533846059}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6829","nonce":2533846059}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6832","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6833","nonce":2533846059}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6831","nonce":2533846059}]},"public_addr":"192.168.123.106:6827/2533846059","cluster_addr":"192.168.123.106:6829/2533846059","heartbeat_back_addr":"192.168.123.106:6833/2533846059","heartbeat_front_addr":"192.168.123.106:6831/2533846059","state":["exists","up"]},{"osd":7,"uuid":"a1f6b136-68c5-4bc2-909e-86f730593029","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6825","nonce":2024432096}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6827","nonce":2024432096}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6831","nonce":2024432096}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6829","nonce":2024432096}]},"public_addr":"192.168.123.109:6825/2024432096","cluster_addr":"192.168.123.109:6827/2024432096","heartbeat_back_addr":"192.168.123.109:6831/2024432096","heartbeat_front_addr":"192.168.123.109:6829/2024432096","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:18.365188+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:18.174922+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:20.692775+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:20.827407+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:23.314833+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:23.974134+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:26.621536+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:26.318666+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.106:0/4006549229":"2026-03-11T10:01:28.812447+0000","192.168.123.106:0/64201105":"2026-03-11T10:01:28.812447+0000","192.168.123.106:6800/426319409":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/3695648630":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/2517368353":"2026-03-11T10:00:59.102034+0000","192.168.123.106:6801/426319409":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/3420003679":"2026-03-11T10:00:59.102034+0000","192.168.123.106:0/2452803408":"2026-03-11T10:01:28.812447+0000","192.168.123.106:0/820830405":"2026-03-11T10:00:45.973613+0000","192.168.123.106:6800/2141042021":"2026-03-11T10:00:59.102034+0000","192.168.123.106:6800/1183985513":"2026-03-11T10:01:28.812447+0000","192.168.123.106:6801/2141042021":"2026-03-11T10:00:59.102034+0000","192.168.123.106:0/4258731819":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/2502688737":"2026-03-11T10:00:59.102034+0000","192.168.123.106:6801/1183985513":"2026-03-11T10:01:28.812447+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T10:02:29.360 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-10T10:02:23.152169+0000', 'flags': 32769, 'flags_names': 'hashpspool,creating', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '21', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 7.889999866485596, 'score_stable': 7.889999866485596, 'optimal_score': 0.3799999952316284, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-10T10:02:29.360 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd pool get .mgr pg_num 2026-03-10T10:02:29.549 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:29.811 INFO:teuthology.orchestra.run.vm06.stdout:pg_num: 1 2026-03-10T10:02:29.843 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:29 vm06 ceph-mon[53281]: osdmap e26: 8 total, 8 up, 8 in 2026-03-10T10:02:29.844 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:29 vm06 ceph-mon[53281]: pgmap v44: 1 pgs: 1 creating+remapped; 0 B data, 211 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:29.844 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:29 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:02:29.844 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:29 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/841559838' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:02:29.995 INFO:tasks.cephadm:Setting up client nodes... 2026-03-10T10:02:29.995 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-10T10:02:30.004 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:29 vm09 ceph-mon[54558]: osdmap e26: 8 total, 8 up, 8 in 2026-03-10T10:02:30.004 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:29 vm09 ceph-mon[54558]: pgmap v44: 1 pgs: 1 creating+remapped; 0 B data, 211 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:30.004 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:29 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:02:30.004 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:29 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/841559838' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:02:30.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 sudo[82803]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T10:02:30.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 sudo[82803]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T10:02:30.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 sudo[82803]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T10:02:30.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 sudo[82803]: pam_unix(sudo:session): session closed for user root 2026-03-10T10:02:30.210 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:30.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 sudo[67854]: ceph : PWD=/ ; USER=root ; COMMAND=/usr/sbin/smartctl -x --json=o /dev/vda 2026-03-10T10:02:30.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 sudo[67854]: pam_systemd(sudo:session): Failed to connect to system bus: No such file or directory 2026-03-10T10:02:30.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 sudo[67854]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=167) 2026-03-10T10:02:30.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 sudo[67854]: pam_unix(sudo:session): session closed for user root 2026-03-10T10:02:30.599 INFO:teuthology.orchestra.run.vm06.stdout:[client.0] 2026-03-10T10:02:30.600 INFO:teuthology.orchestra.run.vm06.stdout: key = AQA27K9pOaM0IxAAL1YTN30RGJYihfsnWtdO3A== 2026-03-10T10:02:30.760 DEBUG:teuthology.orchestra.run.vm06:> set -ex 2026-03-10T10:02:30.761 DEBUG:teuthology.orchestra.run.vm06:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-10T10:02:30.761 DEBUG:teuthology.orchestra.run.vm06:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-10T10:02:30.797 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph auth get-or-create client.1 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-10T10:02:30.984 INFO:teuthology.orchestra.run.vm09.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm09/config 2026-03-10T10:02:31.010 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: osdmap e27: 8 total, 8 up, 8 in 2026-03-10T10:02:31.010 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3635302639' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T10:02:31.010 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:02:31.010 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:02:31.010 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm06"}]: dispatch 2026-03-10T10:02:31.010 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T10:02:31.010 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:02:31.010 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm06"}]: dispatch 2026-03-10T10:02:31.010 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T10:02:31.010 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:02:31.010 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/82103536' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T10:02:31.011 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:30 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/82103536' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: osdmap e27: 8 total, 8 up, 8 in 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3635302639' entity='client.admin' cmd=[{"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"}]: dispatch 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm06"}]: dispatch 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm06"}]: dispatch 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "mon metadata", "id": "vm09"}]: dispatch 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/82103536' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T10:02:31.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:30 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/82103536' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T10:02:31.274 INFO:teuthology.orchestra.run.vm09.stdout:[client.1] 2026-03-10T10:02:31.274 INFO:teuthology.orchestra.run.vm09.stdout: key = AQA37K9pnifQDxAA5t/rhSaTHsjpZbHvZCKDfg== 2026-03-10T10:02:31.440 DEBUG:teuthology.orchestra.run.vm09:> set -ex 2026-03-10T10:02:31.440 DEBUG:teuthology.orchestra.run.vm09:> sudo dd of=/etc/ceph/ceph.client.1.keyring 2026-03-10T10:02:31.440 DEBUG:teuthology.orchestra.run.vm09:> sudo chmod 0644 /etc/ceph/ceph.client.1.keyring 2026-03-10T10:02:31.478 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-10T10:02:31.478 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-10T10:02:31.478 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mgr dump --format=json 2026-03-10T10:02:31.671 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:31.756 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:31 vm06 ceph-mon[53281]: osdmap e28: 8 total, 8 up, 8 in 2026-03-10T10:02:31.756 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:31 vm06 ceph-mon[53281]: pgmap v47: 1 pgs: 1 creating+remapped; 0 B data, 211 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:31.756 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:31 vm06 ceph-mon[53281]: from='client.? 192.168.123.109:0/4226622272' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T10:02:31.756 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:31 vm06 ceph-mon[53281]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T10:02:31.756 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:31 vm06 ceph-mon[53281]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T10:02:31.944 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:32.120 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":19,"flags":0,"active_gid":14215,"active_name":"vm06.dkjjvn","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6800","nonce":381093942},{"type":"v1","addr":"192.168.123.106:6801","nonce":381093942}]},"active_addr":"192.168.123.106:6801/381093942","active_change":"2026-03-10T10:01:28.812546+0000","active_mgr_features":4540701547738038271,"available":true,"standbys":[{"gid":14240,"name":"vm09.jzyfcu","mgr_features":4540701547738038271,"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}]}],"modules":["cephadm","dashboard","iostat","nfs","prometheus","restful"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.25.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:10.4.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"HAproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.7.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.2.5","min":"","max":"","enum_allowed":[],"desc":"Nvme-of container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v2.51.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:devbuilds-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba/SMB container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"SNMP Gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"restful","can_run":true,"error_string":"","module_options":{"enable_auth":{"name":"enable_auth","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_requests":{"name":"max_requests","type":"int","level":"advanced","flags":0,"default_value":"500","min":"","max":"","enum_allowed":[],"desc":"Maximum number of requests to keep in memory. When new request comes in, the oldest request will be removed if the number of requests exceeds the max request number. if un-finished request is removed, error message will be logged in the ceph-mgr log.","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"zabbix","can_run":true,"error_string":"","module_options":{"discovery_interval":{"name":"discovery_interval","type":"uint","level":"advanced","flags":0,"default_value":"100","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"identifier":{"name":"identifier","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_host":{"name":"zabbix_host","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_port":{"name":"zabbix_port","type":"int","level":"advanced","flags":0,"default_value":"10051","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"zabbix_sender":{"name":"zabbix_sender","type":"str","level":"advanced","flags":0,"default_value":"/usr/bin/zabbix_sender","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.106:8443/","prometheus":"http://192.168.123.106:9283/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":5,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":1461709022}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":867798164}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":3643976626}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.106:0","nonce":1493821867}]}]} 2026-03-10T10:02:32.122 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-10T10:02:32.122 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-10T10:02:32.122 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd dump --format=json 2026-03-10T10:02:32.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:31 vm09 ceph-mon[54558]: osdmap e28: 8 total, 8 up, 8 in 2026-03-10T10:02:32.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:31 vm09 ceph-mon[54558]: pgmap v47: 1 pgs: 1 creating+remapped; 0 B data, 211 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:32.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:31 vm09 ceph-mon[54558]: from='client.? 192.168.123.109:0/4226622272' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T10:02:32.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:31 vm09 ceph-mon[54558]: from='client.? ' entity='client.admin' cmd=[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]: dispatch 2026-03-10T10:02:32.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:31 vm09 ceph-mon[54558]: from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-10T10:02:32.316 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:32.571 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:32.571 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":28,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","created":"2026-03-10T10:00:34.057913+0000","modified":"2026-03-10T10:02:30.739520+0000","last_up_change":"2026-03-10T10:02:27.714365+0000","last_in_change":"2026-03-10T10:02:08.475060+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":11,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T10:02:23.152169+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"28","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"c6441cba-6366-4f4b-a30e-5f30a7196eb8","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6801","nonce":1572520565}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6803","nonce":1572520565}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6807","nonce":1572520565}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6805","nonce":1572520565}]},"public_addr":"192.168.123.109:6801/1572520565","cluster_addr":"192.168.123.109:6803/1572520565","heartbeat_back_addr":"192.168.123.109:6807/1572520565","heartbeat_front_addr":"192.168.123.109:6805/1572520565","state":["exists","up"]},{"osd":1,"uuid":"984ffa57-5df4-46a4-acfb-9d25f1ce2c76","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6803","nonce":1902557501}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6805","nonce":1902557501}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6809","nonce":1902557501}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6807","nonce":1902557501}]},"public_addr":"192.168.123.106:6803/1902557501","cluster_addr":"192.168.123.106:6805/1902557501","heartbeat_back_addr":"192.168.123.106:6809/1902557501","heartbeat_front_addr":"192.168.123.106:6807/1902557501","state":["exists","up"]},{"osd":2,"uuid":"b830bc44-e908-40b0-aa15-2527e834436e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6811","nonce":4167697000}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6813","nonce":4167697000}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6817","nonce":4167697000}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6815","nonce":4167697000}]},"public_addr":"192.168.123.106:6811/4167697000","cluster_addr":"192.168.123.106:6813/4167697000","heartbeat_back_addr":"192.168.123.106:6817/4167697000","heartbeat_front_addr":"192.168.123.106:6815/4167697000","state":["exists","up"]},{"osd":3,"uuid":"51747f43-8d7c-498e-bab3-fd72ea514f53","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":23,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6809","nonce":3783651379}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6811","nonce":3783651379}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6815","nonce":3783651379}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6813","nonce":3783651379}]},"public_addr":"192.168.123.109:6809/3783651379","cluster_addr":"192.168.123.109:6811/3783651379","heartbeat_back_addr":"192.168.123.109:6815/3783651379","heartbeat_front_addr":"192.168.123.109:6813/3783651379","state":["exists","up"]},{"osd":4,"uuid":"3937fe45-31f7-4447-9ada-32834c16699a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":22,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6819","nonce":1691146704}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6821","nonce":1691146704}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6825","nonce":1691146704}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6823","nonce":1691146704}]},"public_addr":"192.168.123.106:6819/1691146704","cluster_addr":"192.168.123.106:6821/1691146704","heartbeat_back_addr":"192.168.123.106:6825/1691146704","heartbeat_front_addr":"192.168.123.106:6823/1691146704","state":["exists","up"]},{"osd":5,"uuid":"e0feac48-24ef-487a-af9a-a863a28ad6e1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6817","nonce":87938156}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6819","nonce":87938156}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6823","nonce":87938156}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6821","nonce":87938156}]},"public_addr":"192.168.123.109:6817/87938156","cluster_addr":"192.168.123.109:6819/87938156","heartbeat_back_addr":"192.168.123.109:6823/87938156","heartbeat_front_addr":"192.168.123.109:6821/87938156","state":["exists","up"]},{"osd":6,"uuid":"c9a065b3-6872-43ce-a26f-ba45448813d8","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6827","nonce":2533846059}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6829","nonce":2533846059}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6832","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6833","nonce":2533846059}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6831","nonce":2533846059}]},"public_addr":"192.168.123.106:6827/2533846059","cluster_addr":"192.168.123.106:6829/2533846059","heartbeat_back_addr":"192.168.123.106:6833/2533846059","heartbeat_front_addr":"192.168.123.106:6831/2533846059","state":["exists","up"]},{"osd":7,"uuid":"a1f6b136-68c5-4bc2-909e-86f730593029","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":26,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6825","nonce":2024432096}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6827","nonce":2024432096}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6831","nonce":2024432096}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6829","nonce":2024432096}]},"public_addr":"192.168.123.109:6825/2024432096","cluster_addr":"192.168.123.109:6827/2024432096","heartbeat_back_addr":"192.168.123.109:6831/2024432096","heartbeat_front_addr":"192.168.123.109:6829/2024432096","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:18.365188+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:18.174922+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:20.692775+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:20.827407+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:23.314833+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:23.974134+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:26.621536+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:26.318666+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.106:0/4006549229":"2026-03-11T10:01:28.812447+0000","192.168.123.106:0/64201105":"2026-03-11T10:01:28.812447+0000","192.168.123.106:6800/426319409":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/3695648630":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/2517368353":"2026-03-11T10:00:59.102034+0000","192.168.123.106:6801/426319409":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/3420003679":"2026-03-11T10:00:59.102034+0000","192.168.123.106:0/2452803408":"2026-03-11T10:01:28.812447+0000","192.168.123.106:0/820830405":"2026-03-11T10:00:45.973613+0000","192.168.123.106:6800/2141042021":"2026-03-11T10:00:59.102034+0000","192.168.123.106:6800/1183985513":"2026-03-11T10:01:28.812447+0000","192.168.123.106:6801/2141042021":"2026-03-11T10:00:59.102034+0000","192.168.123.106:0/4258731819":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/2502688737":"2026-03-11T10:00:59.102034+0000","192.168.123.106:6801/1183985513":"2026-03-11T10:01:28.812447+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T10:02:32.756 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-10T10:02:32.757 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd dump --format=json 2026-03-10T10:02:32.952 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:32.980 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:32 vm06 ceph-mon[53281]: mgrmap e19: vm06.dkjjvn(active, since 62s), standbys: vm09.jzyfcu 2026-03-10T10:02:32.980 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:32 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/2936224627' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T10:02:32.980 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:32 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1744834432' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:02:33.207 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:33.207 INFO:teuthology.orchestra.run.vm06.stdout:{"epoch":28,"fsid":"ddb3da90-1c67-11f1-bef5-37be0d23c456","created":"2026-03-10T10:00:34.057913+0000","modified":"2026-03-10T10:02:30.739520+0000","last_up_change":"2026-03-10T10:02:27.714365+0000","last_in_change":"2026-03-10T10:02:08.475060+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":11,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"squid","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-10T10:02:23.152169+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"28","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"c6441cba-6366-4f4b-a30e-5f30a7196eb8","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6800","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6801","nonce":1572520565}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6802","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6803","nonce":1572520565}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6806","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6807","nonce":1572520565}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6804","nonce":1572520565},{"type":"v1","addr":"192.168.123.109:6805","nonce":1572520565}]},"public_addr":"192.168.123.109:6801/1572520565","cluster_addr":"192.168.123.109:6803/1572520565","heartbeat_back_addr":"192.168.123.109:6807/1572520565","heartbeat_front_addr":"192.168.123.109:6805/1572520565","state":["exists","up"]},{"osd":1,"uuid":"984ffa57-5df4-46a4-acfb-9d25f1ce2c76","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":17,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6802","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6803","nonce":1902557501}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6804","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6805","nonce":1902557501}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6808","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6809","nonce":1902557501}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6806","nonce":1902557501},{"type":"v1","addr":"192.168.123.106:6807","nonce":1902557501}]},"public_addr":"192.168.123.106:6803/1902557501","cluster_addr":"192.168.123.106:6805/1902557501","heartbeat_back_addr":"192.168.123.106:6809/1902557501","heartbeat_front_addr":"192.168.123.106:6807/1902557501","state":["exists","up"]},{"osd":2,"uuid":"b830bc44-e908-40b0-aa15-2527e834436e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":19,"up_thru":25,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6810","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6811","nonce":4167697000}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6812","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6813","nonce":4167697000}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6816","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6817","nonce":4167697000}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6814","nonce":4167697000},{"type":"v1","addr":"192.168.123.106:6815","nonce":4167697000}]},"public_addr":"192.168.123.106:6811/4167697000","cluster_addr":"192.168.123.106:6813/4167697000","heartbeat_back_addr":"192.168.123.106:6817/4167697000","heartbeat_front_addr":"192.168.123.106:6815/4167697000","state":["exists","up"]},{"osd":3,"uuid":"51747f43-8d7c-498e-bab3-fd72ea514f53","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":20,"up_thru":23,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6808","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6809","nonce":3783651379}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6810","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6811","nonce":3783651379}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6814","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6815","nonce":3783651379}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6812","nonce":3783651379},{"type":"v1","addr":"192.168.123.109:6813","nonce":3783651379}]},"public_addr":"192.168.123.109:6809/3783651379","cluster_addr":"192.168.123.109:6811/3783651379","heartbeat_back_addr":"192.168.123.109:6815/3783651379","heartbeat_front_addr":"192.168.123.109:6813/3783651379","state":["exists","up"]},{"osd":4,"uuid":"3937fe45-31f7-4447-9ada-32834c16699a","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":22,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6818","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6819","nonce":1691146704}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6820","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6821","nonce":1691146704}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6824","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6825","nonce":1691146704}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6822","nonce":1691146704},{"type":"v1","addr":"192.168.123.106:6823","nonce":1691146704}]},"public_addr":"192.168.123.106:6819/1691146704","cluster_addr":"192.168.123.106:6821/1691146704","heartbeat_back_addr":"192.168.123.106:6825/1691146704","heartbeat_front_addr":"192.168.123.106:6823/1691146704","state":["exists","up"]},{"osd":5,"uuid":"e0feac48-24ef-487a-af9a-a863a28ad6e1","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":23,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6816","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6817","nonce":87938156}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6818","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6819","nonce":87938156}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6822","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6823","nonce":87938156}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6820","nonce":87938156},{"type":"v1","addr":"192.168.123.109:6821","nonce":87938156}]},"public_addr":"192.168.123.109:6817/87938156","cluster_addr":"192.168.123.109:6819/87938156","heartbeat_back_addr":"192.168.123.109:6823/87938156","heartbeat_front_addr":"192.168.123.109:6821/87938156","state":["exists","up"]},{"osd":6,"uuid":"c9a065b3-6872-43ce-a26f-ba45448813d8","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6826","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6827","nonce":2533846059}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6828","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6829","nonce":2533846059}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6832","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6833","nonce":2533846059}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.106:6830","nonce":2533846059},{"type":"v1","addr":"192.168.123.106:6831","nonce":2533846059}]},"public_addr":"192.168.123.106:6827/2533846059","cluster_addr":"192.168.123.106:6829/2533846059","heartbeat_back_addr":"192.168.123.106:6833/2533846059","heartbeat_front_addr":"192.168.123.106:6831/2533846059","state":["exists","up"]},{"osd":7,"uuid":"a1f6b136-68c5-4bc2-909e-86f730593029","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":25,"up_thru":26,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6824","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6825","nonce":2024432096}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6826","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6827","nonce":2024432096}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6830","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6831","nonce":2024432096}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.109:6828","nonce":2024432096},{"type":"v1","addr":"192.168.123.109:6829","nonce":2024432096}]},"public_addr":"192.168.123.109:6825/2024432096","cluster_addr":"192.168.123.109:6827/2024432096","heartbeat_back_addr":"192.168.123.109:6831/2024432096","heartbeat_front_addr":"192.168.123.109:6829/2024432096","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:18.365188+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:18.174922+0000","dead_epoch":0},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:20.692775+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:20.827407+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:23.314833+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:23.974134+0000","dead_epoch":0},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:26.621536+0000","dead_epoch":0},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4540701547738038271,"old_weight":0,"last_purged_snaps_scrub":"2026-03-10T10:02:26.318666+0000","dead_epoch":0}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.106:0/4006549229":"2026-03-11T10:01:28.812447+0000","192.168.123.106:0/64201105":"2026-03-11T10:01:28.812447+0000","192.168.123.106:6800/426319409":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/3695648630":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/2517368353":"2026-03-11T10:00:59.102034+0000","192.168.123.106:6801/426319409":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/3420003679":"2026-03-11T10:00:59.102034+0000","192.168.123.106:0/2452803408":"2026-03-11T10:01:28.812447+0000","192.168.123.106:0/820830405":"2026-03-11T10:00:45.973613+0000","192.168.123.106:6800/2141042021":"2026-03-11T10:00:59.102034+0000","192.168.123.106:6800/1183985513":"2026-03-11T10:01:28.812447+0000","192.168.123.106:6801/2141042021":"2026-03-11T10:00:59.102034+0000","192.168.123.106:0/4258731819":"2026-03-11T10:00:45.973613+0000","192.168.123.106:0/2502688737":"2026-03-11T10:00:59.102034+0000","192.168.123.106:6801/1183985513":"2026-03-11T10:01:28.812447+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"jerasure","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-10T10:02:33.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:32 vm09 ceph-mon[54558]: mgrmap e19: vm06.dkjjvn(active, since 62s), standbys: vm09.jzyfcu 2026-03-10T10:02:33.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:32 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/2936224627' entity='client.admin' cmd=[{"prefix": "mgr dump", "format": "json"}]: dispatch 2026-03-10T10:02:33.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:32 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1744834432' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:02:33.377 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph tell osd.0 flush_pg_stats 2026-03-10T10:02:33.377 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph tell osd.1 flush_pg_stats 2026-03-10T10:02:33.377 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph tell osd.2 flush_pg_stats 2026-03-10T10:02:33.377 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph tell osd.3 flush_pg_stats 2026-03-10T10:02:33.377 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph tell osd.4 flush_pg_stats 2026-03-10T10:02:33.378 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph tell osd.5 flush_pg_stats 2026-03-10T10:02:33.378 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph tell osd.6 flush_pg_stats 2026-03-10T10:02:33.378 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph tell osd.7 flush_pg_stats 2026-03-10T10:02:33.772 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:33 vm06 ceph-mon[53281]: pgmap v48: 1 pgs: 1 creating+remapped; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:33.773 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:33 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3632391091' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:02:34.121 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:34.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:33 vm09 ceph-mon[54558]: pgmap v48: 1 pgs: 1 creating+remapped; 0 B data, 212 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:34.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:33 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3632391091' entity='client.admin' cmd=[{"prefix": "osd dump", "format": "json"}]: dispatch 2026-03-10T10:02:34.255 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:34.259 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:34.262 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:34.262 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:34.275 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:34.282 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:34.404 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:34.778 INFO:teuthology.orchestra.run.vm06.stdout:73014444037 2026-03-10T10:02:34.779 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd last-stat-seq osd.0 2026-03-10T10:02:35.009 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:35.344 INFO:teuthology.orchestra.run.vm06.stdout:73014444036 2026-03-10T10:02:35.345 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd last-stat-seq osd.1 2026-03-10T10:02:35.509 INFO:teuthology.orchestra.run.vm06.stdout:81604378628 2026-03-10T10:02:35.509 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd last-stat-seq osd.2 2026-03-10T10:02:35.531 INFO:teuthology.orchestra.run.vm06.stdout:94489280515 2026-03-10T10:02:35.532 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd last-stat-seq osd.4 2026-03-10T10:02:35.542 INFO:teuthology.orchestra.run.vm06.stdout:107374182403 2026-03-10T10:02:35.542 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd last-stat-seq osd.6 2026-03-10T10:02:35.584 INFO:teuthology.orchestra.run.vm06.stdout:98784247811 2026-03-10T10:02:35.585 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd last-stat-seq osd.5 2026-03-10T10:02:35.609 INFO:teuthology.orchestra.run.vm06.stdout:73014444037 2026-03-10T10:02:35.693 INFO:teuthology.orchestra.run.vm06.stdout:107374182403 2026-03-10T10:02:35.693 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd last-stat-seq osd.7 2026-03-10T10:02:35.708 INFO:teuthology.orchestra.run.vm06.stdout:85899345924 2026-03-10T10:02:35.708 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd last-stat-seq osd.3 2026-03-10T10:02:35.865 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444037 got 73014444037 for osd.0 2026-03-10T10:02:35.865 DEBUG:teuthology.parallel:result is None 2026-03-10T10:02:36.218 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:36 vm06 ceph-mon[53281]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:36.218 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:36 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/658322385' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T10:02:36.246 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:36.355 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:36.361 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:36.483 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:36.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:36 vm09 ceph-mon[54558]: pgmap v49: 1 pgs: 1 active+clean; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:36.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:36 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/658322385' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 0}]: dispatch 2026-03-10T10:02:36.496 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:36.758 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:36.816 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:36.942 INFO:teuthology.orchestra.run.vm06.stdout:73014444036 2026-03-10T10:02:37.151 INFO:teuthology.orchestra.run.vm06.stdout:81604378628 2026-03-10T10:02:37.334 INFO:teuthology.orchestra.run.vm06.stdout:107374182402 2026-03-10T10:02:37.335 INFO:teuthology.orchestra.run.vm06.stdout:98784247810 2026-03-10T10:02:37.351 INFO:teuthology.orchestra.run.vm06.stdout:94489280514 2026-03-10T10:02:37.388 INFO:teuthology.orchestra.run.vm06.stdout:85899345924 2026-03-10T10:02:37.491 INFO:tasks.cephadm.ceph_manager.ceph:need seq 81604378628 got 81604378628 for osd.2 2026-03-10T10:02:37.491 DEBUG:teuthology.parallel:result is None 2026-03-10T10:02:37.587 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:37 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1885730678' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T10:02:37.588 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:37 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/2612935085' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T10:02:37.609 INFO:tasks.cephadm.ceph_manager.ceph:need seq 98784247811 got 98784247810 for osd.5 2026-03-10T10:02:37.626 INFO:teuthology.orchestra.run.vm06.stdout:107374182403 2026-03-10T10:02:37.629 INFO:tasks.cephadm.ceph_manager.ceph:need seq 73014444036 got 73014444036 for osd.1 2026-03-10T10:02:37.629 DEBUG:teuthology.parallel:result is None 2026-03-10T10:02:37.642 INFO:tasks.cephadm.ceph_manager.ceph:need seq 107374182403 got 107374182402 for osd.6 2026-03-10T10:02:37.666 INFO:tasks.cephadm.ceph_manager.ceph:need seq 94489280515 got 94489280514 for osd.4 2026-03-10T10:02:37.673 INFO:tasks.cephadm.ceph_manager.ceph:need seq 85899345924 got 85899345924 for osd.3 2026-03-10T10:02:37.673 DEBUG:teuthology.parallel:result is None 2026-03-10T10:02:37.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:37 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1885730678' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 1}]: dispatch 2026-03-10T10:02:37.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:37 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/2612935085' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 2}]: dispatch 2026-03-10T10:02:37.807 INFO:tasks.cephadm.ceph_manager.ceph:need seq 107374182403 got 107374182403 for osd.7 2026-03-10T10:02:37.807 DEBUG:teuthology.parallel:result is None 2026-03-10T10:02:38.610 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd last-stat-seq osd.5 2026-03-10T10:02:38.635 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:38 vm06 ceph-mon[53281]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:38.635 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:38 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/857963654' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T10:02:38.635 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:38 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1106532172' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T10:02:38.635 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:38 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/486936998' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T10:02:38.635 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:38 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/4163691802' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T10:02:38.635 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:38 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/2290047732' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T10:02:38.643 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd last-stat-seq osd.6 2026-03-10T10:02:38.667 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph osd last-stat-seq osd.4 2026-03-10T10:02:38.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:38 vm09 ceph-mon[54558]: pgmap v50: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:38.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:38 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/857963654' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T10:02:38.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:38 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1106532172' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T10:02:38.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:38 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/486936998' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T10:02:38.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:38 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/4163691802' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 3}]: dispatch 2026-03-10T10:02:38.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:38 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/2290047732' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 7}]: dispatch 2026-03-10T10:02:38.850 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:38.898 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:39.092 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:39.250 INFO:teuthology.orchestra.run.vm06.stdout:98784247812 2026-03-10T10:02:39.302 INFO:teuthology.orchestra.run.vm06.stdout:107374182403 2026-03-10T10:02:39.415 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:39 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1348432530' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T10:02:39.415 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:39 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/1699417124' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T10:02:39.433 INFO:tasks.cephadm.ceph_manager.ceph:need seq 98784247811 got 98784247812 for osd.5 2026-03-10T10:02:39.433 DEBUG:teuthology.parallel:result is None 2026-03-10T10:02:39.469 INFO:teuthology.orchestra.run.vm06.stdout:94489280516 2026-03-10T10:02:39.498 INFO:tasks.cephadm.ceph_manager.ceph:need seq 107374182403 got 107374182403 for osd.6 2026-03-10T10:02:39.498 DEBUG:teuthology.parallel:result is None 2026-03-10T10:02:39.647 INFO:tasks.cephadm.ceph_manager.ceph:need seq 94489280515 got 94489280516 for osd.4 2026-03-10T10:02:39.647 DEBUG:teuthology.parallel:result is None 2026-03-10T10:02:39.647 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-10T10:02:39.647 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph pg dump --format=json 2026-03-10T10:02:39.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:39 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1348432530' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 5}]: dispatch 2026-03-10T10:02:39.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:39 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/1699417124' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 6}]: dispatch 2026-03-10T10:02:39.827 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:40.075 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:40.075 INFO:teuthology.orchestra.run.vm06.stderr:dumped all 2026-03-10T10:02:40.236 INFO:teuthology.orchestra.run.vm06.stdout:{"pg_ready":true,"pg_map":{"version":51,"stamp":"2026-03-10T10:02:39.126961+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":3,"kb":167739392,"kb_used":218624,"kb_used_data":3500,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167520768,"statfs":{"total":171765137408,"available":171541266432,"internally_reserved":0,"allocated":3584000,"data_stored":2198520,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12711,"internal_metadata":219663961},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.000528"},"pg_stats":[{"pgid":"1.0","version":"27'32","reported_seq":57,"reported_epoch":28,"state":"active+clean","last_fresh":"2026-03-10T10:02:31.405519+0000","last_change":"2026-03-10T10:02:29.745548+0000","last_active":"2026-03-10T10:02:31.405519+0000","last_peered":"2026-03-10T10:02:31.405519+0000","last_clean":"2026-03-10T10:02:31.405519+0000","last_became_active":"2026-03-10T10:02:29.744680+0000","last_became_peered":"2026-03-10T10:02:29.744680+0000","last_unstale":"2026-03-10T10:02:31.405519+0000","last_undegraded":"2026-03-10T10:02:31.405519+0000","last_fullsized":"2026-03-10T10:02:31.405519+0000","mapping_epoch":26,"log_start":"0'0","ondisk_log_start":"0'0","created":20,"last_epoch_clean":27,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T10:02:23.285784+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T10:02:23.285784+0000","last_clean_scrub_stamp":"2026-03-10T10:02:23.285784+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:21:32.089982+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,2,3],"acting":[7,2,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":7,"up_from":25,"seq":107374182404,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27696,"kb_used_data":720,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939728,"statfs":{"total":21470642176,"available":21442281472,"internally_reserved":0,"allocated":737280,"data_stored":561865,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":25,"seq":107374182404,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27112,"kb_used_data":268,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940312,"statfs":{"total":21470642176,"available":21442879488,"internally_reserved":0,"allocated":274432,"data_stored":102585,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":23,"seq":98784247812,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27108,"kb_used_data":268,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940316,"statfs":{"total":21470642176,"available":21442883584,"internally_reserved":0,"allocated":274432,"data_stored":102585,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":22,"seq":94489280516,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27112,"kb_used_data":268,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940312,"statfs":{"total":21470642176,"available":21442879488,"internally_reserved":0,"allocated":274432,"data_stored":102585,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":20,"seq":85899345924,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27688,"kb_used_data":720,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939736,"statfs":{"total":21470642176,"available":21442289664,"internally_reserved":0,"allocated":737280,"data_stored":561865,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":19,"seq":81604378629,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27692,"kb_used_data":720,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939732,"statfs":{"total":21470642176,"available":21442285568,"internally_reserved":0,"allocated":737280,"data_stored":561865,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":17,"seq":73014444037,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27108,"kb_used_data":268,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940316,"statfs":{"total":21470642176,"available":21442883584,"internally_reserved":0,"allocated":274432,"data_stored":102585,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":17,"seq":73014444038,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27108,"kb_used_data":268,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940316,"statfs":{"total":21470642176,"available":21442883584,"internally_reserved":0,"allocated":274432,"data_stored":102585,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T10:02:40.236 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph pg dump --format=json 2026-03-10T10:02:40.433 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:40.466 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:40 vm06 ceph-mon[53281]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:40.466 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:40 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/3378338837' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T10:02:40.680 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:40.680 INFO:teuthology.orchestra.run.vm06.stderr:dumped all 2026-03-10T10:02:40.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:40 vm09 ceph-mon[54558]: pgmap v51: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:40.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:40 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/3378338837' entity='client.admin' cmd=[{"prefix": "osd last-stat-seq", "id": 4}]: dispatch 2026-03-10T10:02:40.840 INFO:teuthology.orchestra.run.vm06.stdout:{"pg_ready":true,"pg_map":{"version":51,"stamp":"2026-03-10T10:02:39.126961+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":3,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":3,"kb":167739392,"kb_used":218624,"kb_used_data":3500,"kb_used_omap":12,"kb_used_meta":214515,"kb_avail":167520768,"statfs":{"total":171765137408,"available":171541266432,"internally_reserved":0,"allocated":3584000,"data_stored":2198520,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":12711,"internal_metadata":219663961},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":0,"num_objects":0,"num_object_clones":0,"num_object_copies":0,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":0,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.000528"},"pg_stats":[{"pgid":"1.0","version":"27'32","reported_seq":57,"reported_epoch":28,"state":"active+clean","last_fresh":"2026-03-10T10:02:31.405519+0000","last_change":"2026-03-10T10:02:29.745548+0000","last_active":"2026-03-10T10:02:31.405519+0000","last_peered":"2026-03-10T10:02:31.405519+0000","last_clean":"2026-03-10T10:02:31.405519+0000","last_became_active":"2026-03-10T10:02:29.744680+0000","last_became_peered":"2026-03-10T10:02:29.744680+0000","last_unstale":"2026-03-10T10:02:31.405519+0000","last_undegraded":"2026-03-10T10:02:31.405519+0000","last_fullsized":"2026-03-10T10:02:31.405519+0000","mapping_epoch":26,"log_start":"0'0","ondisk_log_start":"0'0","created":20,"last_epoch_clean":27,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-10T10:02:23.285784+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-10T10:02:23.285784+0000","last_clean_scrub_stamp":"2026-03-10T10:02:23.285784+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-11T11:21:32.089982+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[7,2,3],"acting":[7,2,3],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":7,"acting_primary":7,"purged_snaps":[]}],"pool_stats":[{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":1388544,"data_stored":1377840,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":3}],"osd_stats":[{"osd":7,"up_from":25,"seq":107374182404,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27696,"kb_used_data":720,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939728,"statfs":{"total":21470642176,"available":21442281472,"internally_reserved":0,"allocated":737280,"data_stored":561865,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1585,"internal_metadata":27457999},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":25,"seq":107374182404,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27112,"kb_used_data":268,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940312,"statfs":{"total":21470642176,"available":21442879488,"internally_reserved":0,"allocated":274432,"data_stored":102585,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":23,"seq":98784247812,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27108,"kb_used_data":268,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940316,"statfs":{"total":21470642176,"available":21442883584,"internally_reserved":0,"allocated":274432,"data_stored":102585,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1589,"internal_metadata":27457995},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":22,"seq":94489280516,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27112,"kb_used_data":268,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940312,"statfs":{"total":21470642176,"available":21442879488,"internally_reserved":0,"allocated":274432,"data_stored":102585,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":20,"seq":85899345924,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27688,"kb_used_data":720,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939736,"statfs":{"total":21470642176,"available":21442289664,"internally_reserved":0,"allocated":737280,"data_stored":561865,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1588,"internal_metadata":27457996},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":19,"seq":81604378629,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27692,"kb_used_data":720,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20939732,"statfs":{"total":21470642176,"available":21442285568,"internally_reserved":0,"allocated":737280,"data_stored":561865,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":17,"seq":73014444037,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27108,"kb_used_data":268,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940316,"statfs":{"total":21470642176,"available":21442883584,"internally_reserved":0,"allocated":274432,"data_stored":102585,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":17,"seq":73014444038,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27108,"kb_used_data":268,"kb_used_omap":1,"kb_used_meta":26814,"kb_avail":20940316,"statfs":{"total":21470642176,"available":21442883584,"internally_reserved":0,"allocated":274432,"data_stored":102585,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":1590,"internal_metadata":27457994},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":7,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-10T10:02:40.840 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-10T10:02:40.840 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-10T10:02:40.840 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-10T10:02:40.840 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph health --format=json 2026-03-10T10:02:41.032 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:41.307 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:41.357 INFO:teuthology.orchestra.run.vm06.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-10T10:02:41.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:41 vm06 ceph-mon[53281]: from='client.14530 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:02:41.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:41 vm06 ceph-mon[53281]: from='client.14534 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:02:41.615 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:41 vm06 ceph-mon[53281]: from='client.? 192.168.123.106:0/763092607' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T10:02:41.643 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-10T10:02:41.643 INFO:tasks.cephadm:Setup complete, yielding 2026-03-10T10:02:41.644 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T10:02:41.646 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm06.local 2026-03-10T10:02:41.646 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch status' 2026-03-10T10:02:41.833 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:41.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:41 vm09 ceph-mon[54558]: from='client.14530 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:02:41.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:41 vm09 ceph-mon[54558]: from='client.14534 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:02:41.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:41 vm09 ceph-mon[54558]: from='client.? 192.168.123.106:0/763092607' entity='client.admin' cmd=[{"prefix": "health", "format": "json"}]: dispatch 2026-03-10T10:02:42.089 INFO:teuthology.orchestra.run.vm06.stdout:Backend: cephadm 2026-03-10T10:02:42.089 INFO:teuthology.orchestra.run.vm06.stdout:Available: Yes 2026-03-10T10:02:42.089 INFO:teuthology.orchestra.run.vm06.stdout:Paused: No 2026-03-10T10:02:42.252 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch ps' 2026-03-10T10:02:42.473 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:alertmanager.vm06 vm06 *:9093,9094 running (53s) 15s ago 90s 19.3M - 0.25.0 c8568f914cd2 2515739b1215 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:ceph-exporter.vm06 vm06 *:9926 running (97s) 15s ago 97s 8883k - 19.2.3-678-ge911bdeb 654f31e6858e a5751c00e778 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:ceph-exporter.vm09 vm09 *:9926 running (68s) 16s ago 68s 6614k - 19.2.3-678-ge911bdeb 654f31e6858e f34a9942447f 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:crash.vm06 vm06 running (96s) 15s ago 96s 7650k - 19.2.3-678-ge911bdeb 654f31e6858e 67ac1dfce5d4 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:crash.vm09 vm09 running (67s) 16s ago 67s 7658k - 19.2.3-678-ge911bdeb 654f31e6858e cb9dacbe6081 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:grafana.vm06 vm06 *:3000 running (51s) 15s ago 83s 72.2M - 10.4.0 c8b91775d855 be57c5781783 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:mgr.vm06.dkjjvn vm06 *:9283,8765,8443 running (2m) 15s ago 2m 542M - 19.2.3-678-ge911bdeb 654f31e6858e 80e5643fbe08 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:mgr.vm09.jzyfcu vm09 *:8443,9283,8765 running (63s) 16s ago 63s 489M - 19.2.3-678-ge911bdeb 654f31e6858e 88a536ef8ea4 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:mon.vm06 vm06 running (2m) 15s ago 2m 48.5M 2048M 19.2.3-678-ge911bdeb 654f31e6858e f501e0d9f8a9 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:mon.vm09 vm09 running (62s) 16s ago 61s 42.9M 2048M 19.2.3-678-ge911bdeb 654f31e6858e 45d247861a5a 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:node-exporter.vm06 vm06 *:9100 running (94s) 15s ago 94s 9311k - 1.7.0 72c9c2088986 dde2f9d5f123 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:node-exporter.vm09 vm09 *:9100 running (64s) 16s ago 64s 9005k - 1.7.0 72c9c2088986 63d7ddb381cf 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:osd.0 vm09 running (27s) 16s ago 27s 32.1M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 09c6b7043ed8 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:osd.1 vm06 running (27s) 15s ago 27s 56.6M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 8125ae25527f 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:osd.2 vm06 running (24s) 15s ago 24s 33.9M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 394b47653578 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:osd.3 vm09 running (24s) 16s ago 24s 53.2M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 96ad4ab42962 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:osd.4 vm06 running (21s) 15s ago 21s 32.3M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 2469329ac060 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:osd.5 vm09 running (21s) 16s ago 21s 58.5M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 26c9ae68ff2f 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:osd.6 vm06 running (18s) 15s ago 18s 25.8M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 4748ed2672ec 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:osd.7 vm09 running (18s) 16s ago 18s 24.5M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 7afa95a15e8f 2026-03-10T10:02:42.729 INFO:teuthology.orchestra.run.vm06.stdout:prometheus.vm06 vm06 *:9095 running (50s) 15s ago 78s 31.7M - 2.51.0 1d3b7f56885b 21766952bb24 2026-03-10T10:02:42.764 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:42 vm06 ceph-mon[53281]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:42.914 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch ls' 2026-03-10T10:02:42.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:42 vm09 ceph-mon[54558]: pgmap v52: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:43.108 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:43.357 INFO:teuthology.orchestra.run.vm06.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-10T10:02:43.357 INFO:teuthology.orchestra.run.vm06.stdout:alertmanager ?:9093,9094 1/1 16s ago 109s count:1 2026-03-10T10:02:43.357 INFO:teuthology.orchestra.run.vm06.stdout:ceph-exporter ?:9926 2/2 17s ago 110s * 2026-03-10T10:02:43.357 INFO:teuthology.orchestra.run.vm06.stdout:crash 2/2 17s ago 111s * 2026-03-10T10:02:43.357 INFO:teuthology.orchestra.run.vm06.stdout:grafana ?:3000 1/1 16s ago 109s count:1 2026-03-10T10:02:43.357 INFO:teuthology.orchestra.run.vm06.stdout:mgr 2/2 17s ago 111s count:2 2026-03-10T10:02:43.357 INFO:teuthology.orchestra.run.vm06.stdout:mon 2/2 17s ago 94s vm06:192.168.123.106=vm06;vm09:192.168.123.109=vm09;count:2 2026-03-10T10:02:43.357 INFO:teuthology.orchestra.run.vm06.stdout:node-exporter ?:9100 2/2 17s ago 109s * 2026-03-10T10:02:43.357 INFO:teuthology.orchestra.run.vm06.stdout:osd.all-available-devices 8 17s ago 53s * 2026-03-10T10:02:43.357 INFO:teuthology.orchestra.run.vm06.stdout:prometheus ?:9095 1/1 16s ago 110s count:1 2026-03-10T10:02:43.555 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch host ls' 2026-03-10T10:02:43.740 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:43.784 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:43 vm06 ceph-mon[53281]: from='client.14542 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:43.984 INFO:teuthology.orchestra.run.vm06.stdout:HOST ADDR LABELS STATUS 2026-03-10T10:02:43.984 INFO:teuthology.orchestra.run.vm06.stdout:vm06 192.168.123.106 2026-03-10T10:02:43.984 INFO:teuthology.orchestra.run.vm06.stdout:vm09 192.168.123.109 2026-03-10T10:02:43.984 INFO:teuthology.orchestra.run.vm06.stdout:2 hosts in cluster 2026-03-10T10:02:43.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:43 vm09 ceph-mon[54558]: from='client.14542 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:44.159 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch device ls' 2026-03-10T10:02:44.348 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:44.640 INFO:teuthology.orchestra.run.vm06.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-10T10:02:44.640 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 16s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T10:02:44.640 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vdb hdd DWNBRSTVMM06001 20.0G No 16s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:44.640 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vdc hdd DWNBRSTVMM06002 20.0G No 16s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:44.640 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vdd hdd DWNBRSTVMM06003 20.0G No 16s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:44.640 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vde hdd DWNBRSTVMM06004 20.0G No 16s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:44.640 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 17s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T10:02:44.640 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vdb hdd DWNBRSTVMM09001 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:44.640 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vdc hdd DWNBRSTVMM09002 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:44.640 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vdd hdd DWNBRSTVMM09003 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:44.640 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vde hdd DWNBRSTVMM09004 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:44.650 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:44 vm06 ceph-mon[53281]: from='client.14546 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:44.650 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:44 vm06 ceph-mon[53281]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:44.650 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:44 vm06 ceph-mon[53281]: from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:44.650 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:44 vm06 ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:02:44.793 INFO:teuthology.run_tasks:Running task vip... 2026-03-10T10:02:44.796 INFO:tasks.vip:Allocating static IPs for each host... 2026-03-10T10:02:44.796 INFO:tasks.vip:peername 192.168.123.106 2026-03-10T10:02:44.796 INFO:tasks.vip:192.168.123.106 in 192.168.123.0/24, pos 105 2026-03-10T10:02:44.796 INFO:tasks.vip:vm06.local static 12.12.0.106, vnet 12.12.0.0/22 2026-03-10T10:02:44.796 INFO:tasks.vip:VIPs are [IPv4Address('12.12.1.106')] 2026-03-10T10:02:44.796 DEBUG:teuthology.orchestra.run.vm06:> sudo ip route ls 2026-03-10T10:02:44.819 INFO:teuthology.orchestra.run.vm06.stdout:default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.106 metric 100 2026-03-10T10:02:44.819 INFO:teuthology.orchestra.run.vm06.stdout:192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.106 metric 100 2026-03-10T10:02:44.820 INFO:tasks.vip:Configuring 12.12.0.106 on vm06.local iface eth0... 2026-03-10T10:02:44.820 DEBUG:teuthology.orchestra.run.vm06:> sudo ip addr add 12.12.0.106/22 dev eth0 2026-03-10T10:02:44.892 INFO:tasks.vip:peername 192.168.123.109 2026-03-10T10:02:44.892 INFO:tasks.vip:192.168.123.109 in 192.168.123.0/24, pos 108 2026-03-10T10:02:44.892 INFO:tasks.vip:vm09.local static 12.12.0.109, vnet 12.12.0.0/22 2026-03-10T10:02:44.892 DEBUG:teuthology.orchestra.run.vm09:> sudo ip route ls 2026-03-10T10:02:44.913 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:44 vm09 ceph-mon[54558]: from='client.14546 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:44.913 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:44 vm09 ceph-mon[54558]: pgmap v53: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:44.913 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:44 vm09 ceph-mon[54558]: from='client.14550 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:44.914 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:44 vm09 ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:02:44.917 INFO:teuthology.orchestra.run.vm09.stdout:default via 192.168.123.1 dev eth0 proto dhcp src 192.168.123.109 metric 100 2026-03-10T10:02:44.917 INFO:teuthology.orchestra.run.vm09.stdout:192.168.123.0/24 dev eth0 proto kernel scope link src 192.168.123.109 metric 100 2026-03-10T10:02:44.918 INFO:tasks.vip:Configuring 12.12.0.109 on vm09.local iface eth0... 2026-03-10T10:02:44.918 DEBUG:teuthology.orchestra.run.vm09:> sudo ip addr add 12.12.0.109/22 dev eth0 2026-03-10T10:02:44.988 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T10:02:44.991 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm06.local 2026-03-10T10:02:44.991 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch device ls --refresh' 2026-03-10T10:02:45.183 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:45.450 INFO:teuthology.orchestra.run.vm06.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-10T10:02:45.450 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 17s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T10:02:45.450 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vdb hdd DWNBRSTVMM06001 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:45.450 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vdc hdd DWNBRSTVMM06002 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:45.450 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vdd hdd DWNBRSTVMM06003 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:45.450 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vde hdd DWNBRSTVMM06004 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:45.450 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 17s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T10:02:45.450 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vdb hdd DWNBRSTVMM09001 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:45.450 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vdc hdd DWNBRSTVMM09002 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:45.450 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vdd hdd DWNBRSTVMM09003 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:45.450 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vde hdd DWNBRSTVMM09004 20.0G No 17s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:02:45.622 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-10T10:02:45.625 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm06.local 2026-03-10T10:02:45.625 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'systemctl stop nfs-server' 2026-03-10T10:02:45.665 INFO:teuthology.orchestra.run.vm06.stderr:+ systemctl stop nfs-server 2026-03-10T10:02:45.674 INFO:tasks.vip:Running commands on role host.b host ubuntu@vm09.local 2026-03-10T10:02:45.674 DEBUG:teuthology.orchestra.run.vm09:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'systemctl stop nfs-server' 2026-03-10T10:02:45.711 INFO:teuthology.orchestra.run.vm09.stderr:+ systemctl stop nfs-server 2026-03-10T10:02:45.719 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T10:02:45.723 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm06.local 2026-03-10T10:02:45.723 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch apply rgw foorgw --port 8800' 2026-03-10T10:02:45.761 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:45 vm06.local ceph-mon[53281]: from='client.14554 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:45.762 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:45 vm06.local ceph-mon[53281]: from='client.14558 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:45.762 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:45 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:02:45.819 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:45 vm09.local ceph-mon[54558]: from='client.14554 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:45.819 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:45 vm09.local ceph-mon[54558]: from='client.14558 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:45.819 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:45 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:02:45.966 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:46.255 INFO:teuthology.orchestra.run.vm06.stdout:Scheduled rgw.foorgw update... 2026-03-10T10:02:46.446 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph nfs cluster create foo --ingress --virtual-ip 12.12.1.106/22' 2026-03-10T10:02:46.729 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:46.784 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:46 vm06.local ceph-mon[53281]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:46.784 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:46 vm06.local ceph-mon[53281]: from='client.14562 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:46.784 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:46 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:46.784 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:46 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:46.784 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:46 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:46.920 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:46 vm09.local ceph-mon[54558]: pgmap v54: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:46.920 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:46 vm09.local ceph-mon[54558]: from='client.14562 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "refresh": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:46.920 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:46 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:46.920 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:46 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:46.920 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:46 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:47.964 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:47 vm09.local ceph-mon[54558]: from='client.14566 -' entity='client.admin' cmd=[{"prefix": "orch apply rgw", "svc_id": "foorgw", "port": 8800, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:47.964 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:47 vm09.local ceph-mon[54558]: Saving service rgw.foorgw spec with placement count:2 2026-03-10T10:02:47.964 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:47 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:47.964 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:47 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:47.964 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:47 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:02:47.964 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:47 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:47.964 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:47 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:47.991 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:47 vm06.local ceph-mon[53281]: from='client.14566 -' entity='client.admin' cmd=[{"prefix": "orch apply rgw", "svc_id": "foorgw", "port": 8800, "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:47.991 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:47 vm06.local ceph-mon[53281]: Saving service rgw.foorgw spec with placement count:2 2026-03-10T10:02:47.991 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:47 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:47.991 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:47 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:47.991 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:47 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]: dispatch 2026-03-10T10:02:47.991 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:47 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:47.991 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:47 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:48.949 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-10T10:02:48.952 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm06.local 2026-03-10T10:02:48.952 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'dnf install -y python3-boto3 || apt install -y python3-boto3' 2026-03-10T10:02:48.982 INFO:teuthology.orchestra.run.vm06.stderr:+ dnf install -y python3-boto3 2026-03-10T10:02:49.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:48 vm06.local ceph-mon[53281]: from='client.14570 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "foo", "ingress": true, "virtual_ip": "12.12.1.106/22", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:49.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:48 vm06.local ceph-mon[53281]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:49.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:48 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:02:49.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:48 vm06.local ceph-mon[53281]: osdmap e29: 8 total, 8 up, 8 in 2026-03-10T10:02:49.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:48 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-10T10:02:49.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:48 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:48 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:48 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:48 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:48 vm09.local ceph-mon[54558]: from='client.14570 -' entity='client.admin' cmd=[{"prefix": "nfs cluster create", "cluster_id": "foo", "ingress": true, "virtual_ip": "12.12.1.106/22", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:02:49.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:48 vm09.local ceph-mon[54558]: pgmap v55: 1 pgs: 1 active+clean; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:49.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:48 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pool create", "pool": ".nfs", "yes_i_really_mean_it": true}]': finished 2026-03-10T10:02:49.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:48 vm09.local ceph-mon[54558]: osdmap e29: 8 total, 8 up, 8 in 2026-03-10T10:02:49.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:48 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]: dispatch 2026-03-10T10:02:49.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:48 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:48 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:48 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:48 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.367 INFO:teuthology.orchestra.run.vm06.stdout:Last metadata expiration check: 0:04:11 ago on Tue 10 Mar 2026 09:58:38 AM UTC. 2026-03-10T10:02:49.466 INFO:teuthology.orchestra.run.vm06.stdout:Dependencies resolved. 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout: Package Arch Version Repository Size 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout:Installing: 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout: python3-boto3 noarch 1.28.62-1.el9 epel 164 k 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout:Installing dependencies: 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout: python3-botocore noarch 1.31.62-2.el9 epel 6.1 M 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath noarch 1.0.1-1.el9 appstream 48 k 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout: python3-s3transfer noarch 0.7.0-1.el9 epel 113 k 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3 noarch 1.26.5-7.el9 baseos 218 k 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout:Transaction Summary 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout:================================================================================ 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout:Install 5 Packages 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout:Total download size: 6.6 M 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout:Installed size: 86 M 2026-03-10T10:02:49.467 INFO:teuthology.orchestra.run.vm06.stdout:Downloading Packages: 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: osdmap e30: 8 total, 8 up, 8 in 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: Saving service nfs.foo spec with placement count:1 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: Saving service ingress.nfs.foo spec with placement count:2 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foorgw.vm09.waheeg", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foorgw.vm09.waheeg", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: Deploying daemon rgw.foorgw.vm09.waheeg on vm09 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: pgmap v58: 33 pgs: 1 active+clean, 4 creating+activating, 6 creating+peering, 22 unknown; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:49.939 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:49 vm09.local ceph-mon[54558]: osdmap e31: 8 total, 8 up, 8 in 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pool application enable", "pool": ".nfs", "app": "nfs"}]': finished 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: osdmap e30: 8 total, 8 up, 8 in 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: Saving service nfs.foo spec with placement count:1 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: Saving service ingress.nfs.foo spec with placement count:2 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foorgw.vm09.waheeg", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foorgw.vm09.waheeg", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: Deploying daemon rgw.foorgw.vm09.waheeg on vm09 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: pgmap v58: 33 pgs: 1 active+clean, 4 creating+activating, 6 creating+peering, 22 unknown; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:50.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:49 vm06.local ceph-mon[53281]: osdmap e31: 8 total, 8 up, 8 in 2026-03-10T10:02:50.973 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:50 vm06.local ceph-mon[53281]: mgrmap e20: vm06.dkjjvn(active, since 80s), standbys: vm09.jzyfcu 2026-03-10T10:02:50.974 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:50 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:50.974 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:50 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:50.974 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:50 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:50.974 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:50 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foorgw.vm06.iyncag", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T10:02:50.974 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:50 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foorgw.vm06.iyncag", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T10:02:50.974 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:50 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:50.974 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:50 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:50.974 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:50 vm06.local ceph-mon[53281]: Deploying daemon rgw.foorgw.vm06.iyncag on vm06 2026-03-10T10:02:51.062 INFO:teuthology.orchestra.run.vm06.stdout:(1/5): python3-boto3-1.28.62-1.el9.noarch.rpm 1.2 MB/s | 164 kB 00:00 2026-03-10T10:02:51.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:50 vm09.local ceph-mon[54558]: mgrmap e20: vm06.dkjjvn(active, since 80s), standbys: vm09.jzyfcu 2026-03-10T10:02:51.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:50 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:51.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:50 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:51.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:50 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:51.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:50 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.rgw.foorgw.vm06.iyncag", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T10:02:51.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:50 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.rgw.foorgw.vm06.iyncag", "caps": ["mon", "allow *", "mgr", "allow rw", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T10:02:51.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:50 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:51.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:50 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:51.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:50 vm09.local ceph-mon[54558]: Deploying daemon rgw.foorgw.vm06.iyncag on vm06 2026-03-10T10:02:51.306 INFO:teuthology.orchestra.run.vm06.stdout:(2/5): python3-botocore-1.31.62-2.el9.noarch.rp 25 MB/s | 6.1 MB 00:00 2026-03-10T10:02:51.307 INFO:teuthology.orchestra.run.vm06.stdout:(3/5): python3-jmespath-1.0.1-1.el9.noarch.rpm 125 kB/s | 48 kB 00:00 2026-03-10T10:02:51.312 INFO:teuthology.orchestra.run.vm06.stdout:(4/5): python3-urllib3-1.26.5-7.el9.noarch.rpm 567 kB/s | 218 kB 00:00 2026-03-10T10:02:51.326 INFO:teuthology.orchestra.run.vm06.stdout:(5/5): python3-s3transfer-0.7.0-1.el9.noarch.rp 5.4 MB/s | 113 kB 00:00 2026-03-10T10:02:51.327 INFO:teuthology.orchestra.run.vm06.stdout:-------------------------------------------------------------------------------- 2026-03-10T10:02:51.327 INFO:teuthology.orchestra.run.vm06.stdout:Total 3.6 MB/s | 6.6 MB 00:01 2026-03-10T10:02:51.351 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction check 2026-03-10T10:02:51.356 INFO:teuthology.orchestra.run.vm06.stdout:Transaction check succeeded. 2026-03-10T10:02:51.356 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction test 2026-03-10T10:02:51.433 INFO:teuthology.orchestra.run.vm06.stdout:Transaction test succeeded. 2026-03-10T10:02:51.433 INFO:teuthology.orchestra.run.vm06.stdout:Running transaction 2026-03-10T10:02:51.517 INFO:teuthology.orchestra.run.vm06.stdout: Preparing : 1/1 2026-03-10T10:02:51.546 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-jmespath-1.0.1-1.el9.noarch 1/5 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: osdmap e32: 8 total, 8 up, 8 in 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='client.? 192.168.123.109:0/2844119728' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: pgmap v61: 65 pgs: 6 active+clean, 4 creating+activating, 6 creating+peering, 49 unknown; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: Saving service rgw.foorgw spec with placement count:2 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: Creating key for client.nfs.foo.0.0.vm06.giwwfx 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm06.giwwfx", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm06.giwwfx", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: Ensuring nfs.foo.0 is in the ganesha grace table 2026-03-10T10:02:52.187 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-10T10:02:52.188 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-10T10:02:52.188 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:52.188 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-10T10:02:52.188 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-10T10:02:52.188 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: Rados config object exists: conf-nfs.foo 2026-03-10T10:02:52.188 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: Creating key for client.nfs.foo.0.0.vm06.giwwfx-rgw 2026-03-10T10:02:52.188 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm06.giwwfx-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T10:02:52.188 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm06.giwwfx-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T10:02:52.188 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: Bind address in nfs.foo.0.0.vm06.giwwfx's ganesha conf is defaulting to empty 2026-03-10T10:02:52.188 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:52.188 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:51 vm06.local ceph-mon[53281]: Deploying daemon nfs.foo.0.0.vm06.giwwfx on vm06 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: osdmap e32: 8 total, 8 up, 8 in 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='client.? 192.168.123.109:0/2844119728' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]: dispatch 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: pgmap v61: 65 pgs: 6 active+clean, 4 creating+activating, 6 creating+peering, 49 unknown; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: Saving service rgw.foorgw spec with placement count:2 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: Creating key for client.nfs.foo.0.0.vm06.giwwfx 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm06.giwwfx", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]: dispatch 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm06.giwwfx", "caps": ["mon", "allow r", "osd", "allow rw pool=.nfs namespace=foo"]}]': finished 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: Ensuring nfs.foo.0 is in the ganesha grace table 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]: dispatch 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.mgr.nfs.grace.nfs.foo", "caps": ["mon", "allow r", "osd", "allow rwx pool .nfs"]}]': finished 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]: dispatch 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth rm", "entity": "client.mgr.nfs.grace.nfs.foo"}]': finished 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: Rados config object exists: conf-nfs.foo 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: Creating key for client.nfs.foo.0.0.vm06.giwwfx-rgw 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm06.giwwfx-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]: dispatch 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "auth get-or-create", "entity": "client.nfs.foo.0.0.vm06.giwwfx-rgw", "caps": ["mon", "allow r", "osd", "allow rwx tag rgw *=*"]}]': finished 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: Bind address in nfs.foo.0.0.vm06.giwwfx's ganesha conf is defaulting to empty 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:02:52.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:51 vm09.local ceph-mon[54558]: Deploying daemon nfs.foo.0.0.vm06.giwwfx on vm06 2026-03-10T10:02:52.410 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-urllib3-1.26.5-7.el9.noarch 2/5 2026-03-10T10:02:52.434 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-botocore-1.31.62-2.el9.noarch 3/5 2026-03-10T10:02:52.474 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-s3transfer-0.7.0-1.el9.noarch 4/5 2026-03-10T10:02:52.497 INFO:teuthology.orchestra.run.vm06.stdout: Installing : python3-boto3-1.28.62-1.el9.noarch 5/5 2026-03-10T10:02:52.793 INFO:teuthology.orchestra.run.vm06.stdout: Running scriptlet: python3-boto3-1.28.62-1.el9.noarch 5/5 2026-03-10T10:02:52.794 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-urllib3-1.26.5-7.el9.noarch 1/5 2026-03-10T10:02:52.794 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-jmespath-1.0.1-1.el9.noarch 2/5 2026-03-10T10:02:52.794 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-boto3-1.28.62-1.el9.noarch 3/5 2026-03-10T10:02:52.794 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-botocore-1.31.62-2.el9.noarch 4/5 2026-03-10T10:02:52.860 INFO:teuthology.orchestra.run.vm06.stdout: Verifying : python3-s3transfer-0.7.0-1.el9.noarch 5/5 2026-03-10T10:02:52.860 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:52.860 INFO:teuthology.orchestra.run.vm06.stdout:Installed: 2026-03-10T10:02:52.860 INFO:teuthology.orchestra.run.vm06.stdout: python3-boto3-1.28.62-1.el9.noarch python3-botocore-1.31.62-2.el9.noarch 2026-03-10T10:02:52.860 INFO:teuthology.orchestra.run.vm06.stdout: python3-jmespath-1.0.1-1.el9.noarch python3-s3transfer-0.7.0-1.el9.noarch 2026-03-10T10:02:52.860 INFO:teuthology.orchestra.run.vm06.stdout: python3-urllib3-1.26.5-7.el9.noarch 2026-03-10T10:02:52.860 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:02:52.860 INFO:teuthology.orchestra.run.vm06.stdout:Complete! 2026-03-10T10:02:52.925 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c '/home/ubuntu/cephtest/cephadm shell radosgw-admin user create --uid foouser --display-name foo > /tmp/user.json' 2026-03-10T10:02:52.993 INFO:teuthology.orchestra.run.vm06.stderr:+ /home/ubuntu/cephtest/cephadm shell radosgw-admin user create --uid foouser --display-name foo 2026-03-10T10:02:53.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:52 vm06.local ceph-mon[53281]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T10:02:53.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:52 vm06.local ceph-mon[53281]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-10T10:02:53.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:52 vm06.local ceph-mon[53281]: osdmap e33: 8 total, 8 up, 8 in 2026-03-10T10:02:53.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:52 vm06.local ceph-mon[53281]: osdmap e34: 8 total, 8 up, 8 in 2026-03-10T10:02:53.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:52 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T10:02:53.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:52 vm06.local ceph-mon[53281]: from='client.? 192.168.123.109:0/1537317153' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T10:02:53.069 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:52 vm06.local ceph-mon[53281]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T10:02:53.173 INFO:teuthology.orchestra.run.vm06.stderr:Inferring fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:02:53.227 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:02:53.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:52 vm09.local ceph-mon[54558]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T10:02:53.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:52 vm09.local ceph-mon[54558]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd='[{"prefix": "osd pool application enable","pool": ".rgw.root","app": "rgw"}]': finished 2026-03-10T10:02:53.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:52 vm09.local ceph-mon[54558]: osdmap e33: 8 total, 8 up, 8 in 2026-03-10T10:02:53.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:52 vm09.local ceph-mon[54558]: osdmap e34: 8 total, 8 up, 8 in 2026-03-10T10:02:53.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:52 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T10:02:53.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:52 vm09.local ceph-mon[54558]: from='client.? 192.168.123.109:0/1537317153' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T10:02:53.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:52 vm09.local ceph-mon[54558]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]: dispatch 2026-03-10T10:02:53.319 INFO:teuthology.orchestra.run.vm06.stderr:Using ceph image with id '654f31e6858e' and tag 'e911bdebe5c8faa3800735d1568fcdca65db60df' created on 2026-02-25 18:57:17 +0000 UTC 2026-03-10T10:02:53.319 INFO:teuthology.orchestra.run.vm06.stderr:quay.ceph.io/ceph-ci/ceph@sha256:8fda260ab1d2d3118a1622f7df75f44f285dfe74e71793626152a711c12bf2cc 2026-03-10T10:02:54.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:54 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:54.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:54 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:54.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:54 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:54.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:54 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:54.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:54 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:54.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:54 vm09.local ceph-mon[54558]: Deploying daemon haproxy.nfs.foo.vm06.meexod on vm06 2026-03-10T10:02:54.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:54 vm09.local ceph-mon[54558]: pgmap v64: 97 pgs: 23 active+clean, 4 creating+activating, 27 creating+peering, 43 unknown; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:54.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:54 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:54.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:54 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:54.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:54 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:54.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:54 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:54.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:54 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:54.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:54 vm06.local ceph-mon[53281]: Deploying daemon haproxy.nfs.foo.vm06.meexod on vm06 2026-03-10T10:02:54.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:54 vm06.local ceph-mon[53281]: pgmap v64: 97 pgs: 23 active+clean, 4 creating+activating, 27 creating+peering, 43 unknown; 449 KiB data, 213 MiB used, 160 GiB / 160 GiB avail 2026-03-10T10:02:55.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:55 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-10T10:02:55.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:55 vm06.local ceph-mon[53281]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-10T10:02:55.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:55 vm06.local ceph-mon[53281]: osdmap e35: 8 total, 8 up, 8 in 2026-03-10T10:02:55.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:55 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-10T10:02:55.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:55 vm09.local ceph-mon[54558]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.log","app": "rgw"}]': finished 2026-03-10T10:02:55.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:55 vm09.local ceph-mon[54558]: osdmap e35: 8 total, 8 up, 8 in 2026-03-10T10:02:56.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:56 vm06.local ceph-mon[53281]: pgmap v66: 97 pgs: 51 active+clean, 33 creating+peering, 13 unknown; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 4.9 KiB/s rd, 2.3 KiB/s wr, 8 op/s 2026-03-10T10:02:56.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:56 vm06.local ceph-mon[53281]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-10T10:02:56.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:56 vm06.local ceph-mon[53281]: Cluster is now healthy 2026-03-10T10:02:56.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:56 vm06.local ceph-mon[53281]: osdmap e36: 8 total, 8 up, 8 in 2026-03-10T10:02:56.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:56 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T10:02:56.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:56 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/2503557288' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T10:02:56.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:56 vm06.local ceph-mon[53281]: from='client.? 192.168.123.109:0/1537317153' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T10:02:56.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:56 vm06.local ceph-mon[53281]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T10:02:56.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:56 vm06.local ceph-mon[53281]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T10:02:56.319 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:56 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:56.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:56 vm09.local ceph-mon[54558]: pgmap v66: 97 pgs: 51 active+clean, 33 creating+peering, 13 unknown; 449 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 4.9 KiB/s rd, 2.3 KiB/s wr, 8 op/s 2026-03-10T10:02:56.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:56 vm09.local ceph-mon[54558]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-10T10:02:56.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:56 vm09.local ceph-mon[54558]: Cluster is now healthy 2026-03-10T10:02:56.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:56 vm09.local ceph-mon[54558]: osdmap e36: 8 total, 8 up, 8 in 2026-03-10T10:02:56.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:56 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T10:02:56.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:56 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/2503557288' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T10:02:56.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:56 vm09.local ceph-mon[54558]: from='client.? 192.168.123.109:0/1537317153' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T10:02:56.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:56 vm09.local ceph-mon[54558]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T10:02:56.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:56 vm09.local ceph-mon[54558]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]: dispatch 2026-03-10T10:02:56.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:56 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:57.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:57 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T10:02:57.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:57 vm09.local ceph-mon[54558]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T10:02:57.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:57 vm09.local ceph-mon[54558]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T10:02:57.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:57 vm09.local ceph-mon[54558]: osdmap e37: 8 total, 8 up, 8 in 2026-03-10T10:02:57.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:57 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T10:02:57.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:57 vm06.local ceph-mon[53281]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T10:02:57.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:57 vm06.local ceph-mon[53281]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.control","app": "rgw"}]': finished 2026-03-10T10:02:57.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:57 vm06.local ceph-mon[53281]: osdmap e37: 8 total, 8 up, 8 in 2026-03-10T10:02:58.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:58 vm09.local ceph-mon[54558]: pgmap v69: 129 pgs: 101 active+clean, 12 creating+peering, 16 unknown; 450 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 9.1 KiB/s rd, 3.7 KiB/s wr, 14 op/s 2026-03-10T10:02:58.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:58 vm09.local ceph-mon[54558]: osdmap e38: 8 total, 8 up, 8 in 2026-03-10T10:02:58.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:58 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T10:02:58.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:58 vm09.local ceph-mon[54558]: from='client.? 192.168.123.109:0/1537317153' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T10:02:58.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:58 vm09.local ceph-mon[54558]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T10:02:58.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:58 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/2503557288' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T10:02:58.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:58 vm09.local ceph-mon[54558]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T10:02:58.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:58 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:58.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:58 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:58.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:58 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:58.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:58 vm09.local ceph-mon[54558]: Deploying daemon haproxy.nfs.foo.vm09.aharkz on vm09 2026-03-10T10:02:58.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:58 vm06.local ceph-mon[53281]: pgmap v69: 129 pgs: 101 active+clean, 12 creating+peering, 16 unknown; 450 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 9.1 KiB/s rd, 3.7 KiB/s wr, 14 op/s 2026-03-10T10:02:58.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:58 vm06.local ceph-mon[53281]: osdmap e38: 8 total, 8 up, 8 in 2026-03-10T10:02:58.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:58 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T10:02:58.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:58 vm06.local ceph-mon[53281]: from='client.? 192.168.123.109:0/1537317153' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T10:02:58.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:58 vm06.local ceph-mon[53281]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T10:02:58.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:58 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/2503557288' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T10:02:58.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:58 vm06.local ceph-mon[53281]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]: dispatch 2026-03-10T10:02:58.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:58 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:58.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:58 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:58.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:58 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:02:58.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:58 vm06.local ceph-mon[53281]: Deploying daemon haproxy.nfs.foo.vm09.aharkz on vm09 2026-03-10T10:02:59.587 INFO:teuthology.run_tasks:Running task python... 2026-03-10T10:02:59.590 INFO:tasks.python:Running python on role host.a host ubuntu@vm06.local 2026-03-10T10:02:59.590 INFO:tasks.python:import boto3 import json with open('/tmp/user.json', 'rt') as f: info = json.loads(f.read()) s3 = boto3.resource( 's3', aws_access_key_id=info['keys'][0]['access_key'], aws_secret_access_key=info['keys'][0]['secret_key'], endpoint_url='http://localhost:8800', ) bucket = s3.Bucket('foobucket') bucket.create() bucket.put_object(Key='myobject', Body='thebody') 2026-03-10T10:02:59.590 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest python3 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: osdmap e39: 8 total, 8 up, 8 in 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/2503557288' entity='client.admin' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='client.? 192.168.123.109:0/1537317153' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T10:02:59.668 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:02:59 vm06.local ceph-mon[53281]: osdmap e40: 8 total, 8 up, 8 in 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.meta","app": "rgw"}]': finished 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: osdmap e39: 8 total, 8 up, 8 in 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/2503557288' entity='client.admin' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='client.? ' entity='client.admin' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='client.? 192.168.123.109:0/1537317153' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: from='client.? ' entity='client.rgw.foorgw.vm09.waheeg' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.meta", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T10:02:59.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:02:59 vm09.local ceph-mon[54558]: osdmap e40: 8 total, 8 up, 8 in 2026-03-10T10:03:00.597 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:00 vm06.local ceph-mon[53281]: pgmap v72: 161 pgs: 134 active+clean, 1 creating+peering, 26 unknown; 450 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 6.0 KiB/s rd, 1.7 KiB/s wr, 8 op/s 2026-03-10T10:03:00.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:00 vm09.local ceph-mon[54558]: pgmap v72: 161 pgs: 134 active+clean, 1 creating+peering, 26 unknown; 450 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 6.0 KiB/s rd, 1.7 KiB/s wr, 8 op/s 2026-03-10T10:03:01.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:01 vm09.local ceph-mon[54558]: osdmap e41: 8 total, 8 up, 8 in 2026-03-10T10:03:01.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:01 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.buckets.index","app": "rgw"}]: dispatch 2026-03-10T10:03:01.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:01 vm06.local ceph-mon[53281]: osdmap e41: 8 total, 8 up, 8 in 2026-03-10T10:03:01.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:01 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.buckets.index","app": "rgw"}]: dispatch 2026-03-10T10:03:02.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:02 vm09.local ceph-mon[54558]: pgmap v75: 193 pgs: 143 active+clean, 1 creating+peering, 49 unknown; 450 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 3.0 KiB/s rd, 511 B/s wr, 5 op/s 2026-03-10T10:03:02.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:02 vm09.local ceph-mon[54558]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T10:03:02.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:02 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.buckets.index","app": "rgw"}]': finished 2026-03-10T10:03:02.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:02 vm09.local ceph-mon[54558]: osdmap e42: 8 total, 8 up, 8 in 2026-03-10T10:03:02.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:02 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.buckets.index", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:03:02.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:02 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:02.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:02 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:02.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:02 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:02.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:02 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:02.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:02 vm06.local ceph-mon[53281]: pgmap v75: 193 pgs: 143 active+clean, 1 creating+peering, 49 unknown; 450 KiB data, 214 MiB used, 160 GiB / 160 GiB avail; 3.0 KiB/s rd, 511 B/s wr, 5 op/s 2026-03-10T10:03:02.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:02 vm06.local ceph-mon[53281]: Health check failed: 1 pool(s) do not have an application enabled (POOL_APP_NOT_ENABLED) 2026-03-10T10:03:02.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:02 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.buckets.index","app": "rgw"}]': finished 2026-03-10T10:03:02.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:02 vm06.local ceph-mon[53281]: osdmap e42: 8 total, 8 up, 8 in 2026-03-10T10:03:02.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:02 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.buckets.index", "var": "pg_autoscale_bias", "val": "4"}]: dispatch 2026-03-10T10:03:02.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:02 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:02.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:02 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:02.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:02 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:02.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:02 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:03.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:03 vm06.local ceph-mon[53281]: 12.12.1.106 is in 12.12.0.0/22 on vm06 interface eth0 2026-03-10T10:03:03.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:03 vm06.local ceph-mon[53281]: 12.12.1.106 is in 12.12.0.0/22 on vm09 interface eth0 2026-03-10T10:03:03.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:03 vm06.local ceph-mon[53281]: Deploying daemon keepalived.nfs.foo.vm06.ulqfeo on vm06 2026-03-10T10:03:03.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:03 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.buckets.index", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T10:03:03.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:03 vm06.local ceph-mon[53281]: osdmap e43: 8 total, 8 up, 8 in 2026-03-10T10:03:03.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:03 vm09.local ceph-mon[54558]: 12.12.1.106 is in 12.12.0.0/22 on vm06 interface eth0 2026-03-10T10:03:03.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:03 vm09.local ceph-mon[54558]: 12.12.1.106 is in 12.12.0.0/22 on vm09 interface eth0 2026-03-10T10:03:03.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:03 vm09.local ceph-mon[54558]: Deploying daemon keepalived.nfs.foo.vm06.ulqfeo on vm06 2026-03-10T10:03:03.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:03 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.buckets.index", "var": "pg_autoscale_bias", "val": "4"}]': finished 2026-03-10T10:03:03.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:03 vm09.local ceph-mon[54558]: osdmap e43: 8 total, 8 up, 8 in 2026-03-10T10:03:04.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:04 vm09.local ceph-mon[54558]: pgmap v78: 193 pgs: 160 active+clean, 18 creating+peering, 15 unknown; 452 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 102 KiB/s rd, 4.7 KiB/s wr, 204 op/s 2026-03-10T10:03:04.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:04 vm09.local ceph-mon[54558]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-10T10:03:04.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:04 vm09.local ceph-mon[54558]: Cluster is now healthy 2026-03-10T10:03:04.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:04 vm09.local ceph-mon[54558]: osdmap e44: 8 total, 8 up, 8 in 2026-03-10T10:03:04.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:04 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.buckets.data","app": "rgw"}]: dispatch 2026-03-10T10:03:05.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:04 vm06.local ceph-mon[53281]: pgmap v78: 193 pgs: 160 active+clean, 18 creating+peering, 15 unknown; 452 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 102 KiB/s rd, 4.7 KiB/s wr, 204 op/s 2026-03-10T10:03:05.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:04 vm06.local ceph-mon[53281]: Health check cleared: POOL_APP_NOT_ENABLED (was: 1 pool(s) do not have an application enabled) 2026-03-10T10:03:05.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:04 vm06.local ceph-mon[53281]: Cluster is now healthy 2026-03-10T10:03:05.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:04 vm06.local ceph-mon[53281]: osdmap e44: 8 total, 8 up, 8 in 2026-03-10T10:03:05.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:04 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool application enable","pool": "default.rgw.buckets.data","app": "rgw"}]: dispatch 2026-03-10T10:03:05.663 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T10:03:05.665 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm06.local 2026-03-10T10:03:05.665 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph nfs export create rgw --bucket foobucket --cluster-id foo --pseudo-path /foobucket' 2026-03-10T10:03:05.938 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:05.972 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:05 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.buckets.data","app": "rgw"}]': finished 2026-03-10T10:03:05.972 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:05 vm06.local ceph-mon[53281]: osdmap e45: 8 total, 8 up, 8 in 2026-03-10T10:03:05.972 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:05 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.buckets.data", "var": "bulk", "val": "true"}]: dispatch 2026-03-10T10:03:05.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:05 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool application enable","pool": "default.rgw.buckets.data","app": "rgw"}]': finished 2026-03-10T10:03:05.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:05 vm09.local ceph-mon[54558]: osdmap e45: 8 total, 8 up, 8 in 2026-03-10T10:03:05.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:05 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd=[{"prefix": "osd pool set", "pool": "default.rgw.buckets.data", "var": "bulk", "val": "true"}]: dispatch 2026-03-10T10:03:06.531 INFO:teuthology.orchestra.run.vm06.stdout:{ 2026-03-10T10:03:06.531 INFO:teuthology.orchestra.run.vm06.stdout: "bind": "/foobucket", 2026-03-10T10:03:06.531 INFO:teuthology.orchestra.run.vm06.stdout: "cluster": "foo", 2026-03-10T10:03:06.531 INFO:teuthology.orchestra.run.vm06.stdout: "mode": "RW", 2026-03-10T10:03:06.531 INFO:teuthology.orchestra.run.vm06.stdout: "path": "foobucket", 2026-03-10T10:03:06.531 INFO:teuthology.orchestra.run.vm06.stdout: "squash": "none" 2026-03-10T10:03:06.531 INFO:teuthology.orchestra.run.vm06.stdout:} 2026-03-10T10:03:06.705 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-10T10:03:06.708 INFO:tasks.cephadm:Waiting for ceph service nfs.foo to start (timeout 300)... 2026-03-10T10:03:06.708 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch ls -f json 2026-03-10T10:03:06.782 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:06 vm06.local ceph-mon[53281]: pgmap v81: 225 pgs: 176 active+clean, 25 creating+peering, 24 unknown; 454 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 228 KiB/s rd, 8.5 KiB/s wr, 459 op/s 2026-03-10T10:03:06.782 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:06 vm06.local ceph-mon[53281]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.buckets.data", "var": "bulk", "val": "true"}]': finished 2026-03-10T10:03:06.782 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:06 vm06.local ceph-mon[53281]: osdmap e46: 8 total, 8 up, 8 in 2026-03-10T10:03:06.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:06 vm09.local ceph-mon[54558]: pgmap v81: 225 pgs: 176 active+clean, 25 creating+peering, 24 unknown; 454 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 228 KiB/s rd, 8.5 KiB/s wr, 459 op/s 2026-03-10T10:03:06.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:06 vm09.local ceph-mon[54558]: from='client.? 192.168.123.106:0/1586486890' entity='client.rgw.foorgw.vm06.iyncag' cmd='[{"prefix": "osd pool set", "pool": "default.rgw.buckets.data", "var": "bulk", "val": "true"}]': finished 2026-03-10T10:03:06.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:06 vm09.local ceph-mon[54558]: osdmap e46: 8 total, 8 up, 8 in 2026-03-10T10:03:07.014 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:07.272 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:03:07.273 INFO:teuthology.orchestra.run.vm06.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:00:54.272754Z", "last_refresh": "2026-03-10T10:02:47.314857Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:01:34.251508Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-10T10:00:52.371303Z", "last_refresh": "2026-03-10T10:02:46.257584Z", "ports": [9926], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:35.268370Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:00:51.930427Z", "last_refresh": "2026-03-10T10:02:46.257641Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-10T10:00:53.376036Z", "last_refresh": "2026-03-10T10:02:47.314884Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:03:02.410194Z service:ingress.nfs.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.foo", "service_name": "ingress.nfs.foo", "service_type": "ingress", "spec": {"backend_service": "nfs.foo", "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.106/22"}, "status": {"created": "2026-03-10T10:02:48.782609Z", "ports": [2049, 9049], "running": 0, "size": 4, "virtual_ip": "12.12.1.106/22"}}, {"events": ["2026-03-10T10:01:39.184549Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:00:51.537793Z", "last_refresh": "2026-03-10T10:02:46.257708Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:40.822470Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06:192.168.123.106=vm06", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:01:08.713712Z", "last_refresh": "2026-03-10T10:02:46.257739Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:02:52.951281Z service:nfs.foo [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "foo", "service_name": "nfs.foo", "service_type": "nfs", "spec": {"port": 12049}, "status": {"created": "2026-03-10T10:02:48.775768Z", "ports": [12049], "running": 0, "size": 1}}, {"events": ["2026-03-10T10:01:38.185414Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:00:53.882721Z", "last_refresh": "2026-03-10T10:02:46.257676Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:50.126156Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-10T10:01:50.119265Z", "last_refresh": "2026-03-10T10:02:46.257769Z", "running": 8, "size": 8}}, {"events": ["2026-03-10T10:01:40.825549Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:00:52.912255Z", "last_refresh": "2026-03-10T10:02:47.314912Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:02:51.482921Z service:rgw.foorgw [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foorgw", "service_name": "rgw.foorgw", "service_type": "rgw", "spec": {"rgw_frontend_port": 8800}, "status": {"created": "2026-03-10T10:02:51.471546Z", "ports": [8800], "running": 0, "size": 2}}] 2026-03-10T10:03:07.431 INFO:tasks.cephadm:nfs.foo has 0/1 2026-03-10T10:03:07.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:07 vm06.local ceph-mon[53281]: from='client.14616 -' entity='client.admin' cmd=[{"prefix": "nfs export create rgw", "bucket": "foobucket", "cluster_id": "foo", "pseudo_path": "/foobucket", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:07.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:07 vm09.local ceph-mon[54558]: from='client.14616 -' entity='client.admin' cmd=[{"prefix": "nfs export create rgw", "bucket": "foobucket", "cluster_id": "foo", "pseudo_path": "/foobucket", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:08.431 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch ls -f json 2026-03-10T10:03:08.669 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:08.732 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:08 vm06.local ceph-mon[53281]: pgmap v83: 225 pgs: 217 active+clean, 8 creating+peering; 455 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 110 KiB/s rd, 4.7 KiB/s wr, 226 op/s 2026-03-10T10:03:08.732 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:08 vm06.local ceph-mon[53281]: from='client.14656 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:08.967 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:03:08.967 INFO:teuthology.orchestra.run.vm06.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:00:54.272754Z", "last_refresh": "2026-03-10T10:02:47.314857Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:01:34.251508Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-10T10:00:52.371303Z", "last_refresh": "2026-03-10T10:02:46.257584Z", "ports": [9926], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:35.268370Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:00:51.930427Z", "last_refresh": "2026-03-10T10:02:46.257641Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-10T10:00:53.376036Z", "last_refresh": "2026-03-10T10:02:47.314884Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:03:02.410194Z service:ingress.nfs.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.foo", "service_name": "ingress.nfs.foo", "service_type": "ingress", "spec": {"backend_service": "nfs.foo", "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.106/22"}, "status": {"created": "2026-03-10T10:02:48.782609Z", "ports": [2049, 9049], "running": 0, "size": 4, "virtual_ip": "12.12.1.106/22"}}, {"events": ["2026-03-10T10:01:39.184549Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:00:51.537793Z", "last_refresh": "2026-03-10T10:02:46.257708Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:40.822470Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06:192.168.123.106=vm06", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:01:08.713712Z", "last_refresh": "2026-03-10T10:02:46.257739Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:02:52.951281Z service:nfs.foo [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "foo", "service_name": "nfs.foo", "service_type": "nfs", "spec": {"port": 12049}, "status": {"created": "2026-03-10T10:02:48.775768Z", "ports": [12049], "running": 0, "size": 1}}, {"events": ["2026-03-10T10:01:38.185414Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:00:53.882721Z", "last_refresh": "2026-03-10T10:02:46.257676Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:50.126156Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-10T10:01:50.119265Z", "last_refresh": "2026-03-10T10:02:46.257769Z", "running": 8, "size": 8}}, {"events": ["2026-03-10T10:01:40.825549Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:00:52.912255Z", "last_refresh": "2026-03-10T10:02:47.314912Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:02:51.482921Z service:rgw.foorgw [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foorgw", "service_name": "rgw.foorgw", "service_type": "rgw", "spec": {"rgw_frontend_port": 8800}, "status": {"created": "2026-03-10T10:02:51.471546Z", "ports": [8800], "running": 0, "size": 2}}] 2026-03-10T10:03:08.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:08 vm09.local ceph-mon[54558]: pgmap v83: 225 pgs: 217 active+clean, 8 creating+peering; 455 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 110 KiB/s rd, 4.7 KiB/s wr, 226 op/s 2026-03-10T10:03:08.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:08 vm09.local ceph-mon[54558]: from='client.14656 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:10.525 INFO:tasks.cephadm:nfs.foo has 0/1 2026-03-10T10:03:10.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:10 vm09.local ceph-mon[54558]: from='client.24393 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:10.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:10 vm09.local ceph-mon[54558]: pgmap v84: 225 pgs: 222 active+clean, 3 creating+peering; 455 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 95 KiB/s rd, 4.0 KiB/s wr, 189 op/s 2026-03-10T10:03:10.755 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:10 vm06.local ceph-mon[53281]: from='client.24393 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:10.755 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:10 vm06.local ceph-mon[53281]: pgmap v84: 225 pgs: 222 active+clean, 3 creating+peering; 455 KiB data, 215 MiB used, 160 GiB / 160 GiB avail; 95 KiB/s rd, 4.0 KiB/s wr, 189 op/s 2026-03-10T10:03:11.526 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch ls -f json 2026-03-10T10:03:11.848 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:12.122 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:03:12.122 INFO:teuthology.orchestra.run.vm06.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:00:54.272754Z", "last_refresh": "2026-03-10T10:02:47.314857Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:01:34.251508Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-10T10:00:52.371303Z", "last_refresh": "2026-03-10T10:02:46.257584Z", "ports": [9926], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:35.268370Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:00:51.930427Z", "last_refresh": "2026-03-10T10:02:46.257641Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-10T10:00:53.376036Z", "last_refresh": "2026-03-10T10:02:47.314884Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:03:11.839175Z service:ingress.nfs.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.foo", "service_name": "ingress.nfs.foo", "service_type": "ingress", "spec": {"backend_service": "nfs.foo", "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.106/22"}, "status": {"created": "2026-03-10T10:02:48.782609Z", "ports": [2049, 9049], "running": 0, "size": 4, "virtual_ip": "12.12.1.106/22"}}, {"events": ["2026-03-10T10:01:39.184549Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:00:51.537793Z", "last_refresh": "2026-03-10T10:02:46.257708Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:40.822470Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06:192.168.123.106=vm06", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:01:08.713712Z", "last_refresh": "2026-03-10T10:02:46.257739Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:02:52.951281Z service:nfs.foo [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "foo", "service_name": "nfs.foo", "service_type": "nfs", "spec": {"port": 12049}, "status": {"created": "2026-03-10T10:02:48.775768Z", "ports": [12049], "running": 0, "size": 1}}, {"events": ["2026-03-10T10:01:38.185414Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:00:53.882721Z", "last_refresh": "2026-03-10T10:02:46.257676Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:50.126156Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-10T10:01:50.119265Z", "last_refresh": "2026-03-10T10:02:46.257769Z", "running": 8, "size": 8}}, {"events": ["2026-03-10T10:01:40.825549Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:00:52.912255Z", "last_refresh": "2026-03-10T10:02:47.314912Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:02:51.482921Z service:rgw.foorgw [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foorgw", "service_name": "rgw.foorgw", "service_type": "rgw", "spec": {"rgw_frontend_port": 8800}, "status": {"created": "2026-03-10T10:02:51.471546Z", "ports": [8800], "running": 0, "size": 2}}] 2026-03-10T10:03:12.283 INFO:tasks.cephadm:nfs.foo has 0/1 2026-03-10T10:03:12.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:12 vm06.local ceph-mon[53281]: pgmap v85: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 2.5 KiB/s wr, 112 op/s 2026-03-10T10:03:12.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:12 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:12.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:12 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:12.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:12 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:12.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:12 vm09.local ceph-mon[54558]: pgmap v85: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 58 KiB/s rd, 2.5 KiB/s wr, 112 op/s 2026-03-10T10:03:12.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:12 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:12.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:12 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:12.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:12 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:13.284 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch ls -f json 2026-03-10T10:03:13.489 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:13.564 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:13 vm06.local ceph-mon[53281]: 12.12.1.106 is in 12.12.0.0/22 on vm09 interface eth0 2026-03-10T10:03:13.564 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:13 vm06.local ceph-mon[53281]: 12.12.1.106 is in 12.12.0.0/22 on vm06 interface eth0 2026-03-10T10:03:13.564 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:13 vm06.local ceph-mon[53281]: Deploying daemon keepalived.nfs.foo.vm09.kzsmdy on vm09 2026-03-10T10:03:13.564 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:13 vm06.local ceph-mon[53281]: from='client.14662 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:13.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:13 vm09.local ceph-mon[54558]: 12.12.1.106 is in 12.12.0.0/22 on vm09 interface eth0 2026-03-10T10:03:13.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:13 vm09.local ceph-mon[54558]: 12.12.1.106 is in 12.12.0.0/22 on vm06 interface eth0 2026-03-10T10:03:13.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:13 vm09.local ceph-mon[54558]: Deploying daemon keepalived.nfs.foo.vm09.kzsmdy on vm09 2026-03-10T10:03:13.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:13 vm09.local ceph-mon[54558]: from='client.14662 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:13.762 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:03:13.762 INFO:teuthology.orchestra.run.vm06.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:00:54.272754Z", "last_refresh": "2026-03-10T10:02:47.314857Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:01:34.251508Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-10T10:00:52.371303Z", "last_refresh": "2026-03-10T10:02:46.257584Z", "ports": [9926], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:35.268370Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:00:51.930427Z", "last_refresh": "2026-03-10T10:02:46.257641Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-10T10:00:53.376036Z", "last_refresh": "2026-03-10T10:02:47.314884Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:03:11.839175Z service:ingress.nfs.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.foo", "service_name": "ingress.nfs.foo", "service_type": "ingress", "spec": {"backend_service": "nfs.foo", "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.106/22"}, "status": {"created": "2026-03-10T10:02:48.782609Z", "ports": [2049, 9049], "running": 0, "size": 4, "virtual_ip": "12.12.1.106/22"}}, {"events": ["2026-03-10T10:01:39.184549Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:00:51.537793Z", "last_refresh": "2026-03-10T10:02:46.257708Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:40.822470Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06:192.168.123.106=vm06", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:01:08.713712Z", "last_refresh": "2026-03-10T10:02:46.257739Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:02:52.951281Z service:nfs.foo [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "foo", "service_name": "nfs.foo", "service_type": "nfs", "spec": {"port": 12049}, "status": {"created": "2026-03-10T10:02:48.775768Z", "ports": [12049], "running": 0, "size": 1}}, {"events": ["2026-03-10T10:01:38.185414Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:00:53.882721Z", "last_refresh": "2026-03-10T10:02:46.257676Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:50.126156Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-10T10:01:50.119265Z", "last_refresh": "2026-03-10T10:02:46.257769Z", "running": 8, "size": 8}}, {"events": ["2026-03-10T10:01:40.825549Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:00:52.912255Z", "last_refresh": "2026-03-10T10:02:47.314912Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:02:51.482921Z service:rgw.foorgw [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foorgw", "service_name": "rgw.foorgw", "service_type": "rgw", "spec": {"rgw_frontend_port": 8800}, "status": {"created": "2026-03-10T10:02:51.471546Z", "ports": [8800], "running": 0, "size": 2}}] 2026-03-10T10:03:13.918 INFO:tasks.cephadm:nfs.foo has 0/1 2026-03-10T10:03:14.919 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch ls -f json 2026-03-10T10:03:15.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:14 vm06.local ceph-mon[53281]: pgmap v86: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 956 B/s wr, 22 op/s 2026-03-10T10:03:15.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:14 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:03:15.118 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:15.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:14 vm09.local ceph-mon[54558]: pgmap v86: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 15 KiB/s rd, 956 B/s wr, 22 op/s 2026-03-10T10:03:15.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:14 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:03:15.363 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:03:15.363 INFO:teuthology.orchestra.run.vm06.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:00:54.272754Z", "last_refresh": "2026-03-10T10:02:47.314857Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:01:34.251508Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-10T10:00:52.371303Z", "last_refresh": "2026-03-10T10:02:46.257584Z", "ports": [9926], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:35.268370Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:00:51.930427Z", "last_refresh": "2026-03-10T10:02:46.257641Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-10T10:00:53.376036Z", "last_refresh": "2026-03-10T10:02:47.314884Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:03:11.839175Z service:ingress.nfs.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.foo", "service_name": "ingress.nfs.foo", "service_type": "ingress", "spec": {"backend_service": "nfs.foo", "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.106/22"}, "status": {"created": "2026-03-10T10:02:48.782609Z", "ports": [2049, 9049], "running": 0, "size": 4, "virtual_ip": "12.12.1.106/22"}}, {"events": ["2026-03-10T10:01:39.184549Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:00:51.537793Z", "last_refresh": "2026-03-10T10:02:46.257708Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:40.822470Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06:192.168.123.106=vm06", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:01:08.713712Z", "last_refresh": "2026-03-10T10:02:46.257739Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:02:52.951281Z service:nfs.foo [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "foo", "service_name": "nfs.foo", "service_type": "nfs", "spec": {"port": 12049}, "status": {"created": "2026-03-10T10:02:48.775768Z", "ports": [12049], "running": 0, "size": 1}}, {"events": ["2026-03-10T10:01:38.185414Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:00:53.882721Z", "last_refresh": "2026-03-10T10:02:46.257676Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:50.126156Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-10T10:01:50.119265Z", "last_refresh": "2026-03-10T10:02:46.257769Z", "running": 8, "size": 8}}, {"events": ["2026-03-10T10:01:40.825549Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:00:52.912255Z", "last_refresh": "2026-03-10T10:02:47.314912Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:02:51.482921Z service:rgw.foorgw [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foorgw", "service_name": "rgw.foorgw", "service_type": "rgw", "spec": {"rgw_frontend_port": 8800}, "status": {"created": "2026-03-10T10:02:51.471546Z", "ports": [8800], "running": 0, "size": 2}}] 2026-03-10T10:03:15.525 INFO:tasks.cephadm:nfs.foo has 0/1 2026-03-10T10:03:16.060 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:15 vm09.local ceph-mon[54558]: from='client.14666 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:16.060 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:15 vm09.local ceph-mon[54558]: pgmap v87: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 13 KiB/s rd, 818 B/s wr, 19 op/s 2026-03-10T10:03:16.060 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:15 vm09.local ceph-mon[54558]: from='client.14670 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:16.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:15 vm06.local ceph-mon[53281]: from='client.14666 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:16.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:15 vm06.local ceph-mon[53281]: pgmap v87: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 13 KiB/s rd, 818 B/s wr, 19 op/s 2026-03-10T10:03:16.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:15 vm06.local ceph-mon[53281]: from='client.14670 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:16.525 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch ls -f json 2026-03-10T10:03:16.753 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:17.028 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:03:17.028 INFO:teuthology.orchestra.run.vm06.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:00:54.272754Z", "last_refresh": "2026-03-10T10:02:47.314857Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:01:34.251508Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-10T10:00:52.371303Z", "last_refresh": "2026-03-10T10:02:46.257584Z", "ports": [9926], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:35.268370Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:00:51.930427Z", "last_refresh": "2026-03-10T10:02:46.257641Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-10T10:00:53.376036Z", "last_refresh": "2026-03-10T10:02:47.314884Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:03:16.570620Z service:ingress.nfs.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.foo", "service_name": "ingress.nfs.foo", "service_type": "ingress", "spec": {"backend_service": "nfs.foo", "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.106/22"}, "status": {"created": "2026-03-10T10:02:48.782609Z", "ports": [2049, 9049], "running": 0, "size": 4, "virtual_ip": "12.12.1.106/22"}}, {"events": ["2026-03-10T10:01:39.184549Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:00:51.537793Z", "last_refresh": "2026-03-10T10:02:46.257708Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:40.822470Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06:192.168.123.106=vm06", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:01:08.713712Z", "last_refresh": "2026-03-10T10:02:46.257739Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:02:52.951281Z service:nfs.foo [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "foo", "service_name": "nfs.foo", "service_type": "nfs", "spec": {"port": 12049}, "status": {"created": "2026-03-10T10:02:48.775768Z", "ports": [12049], "running": 0, "size": 1}}, {"events": ["2026-03-10T10:01:38.185414Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:00:53.882721Z", "last_refresh": "2026-03-10T10:02:46.257676Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:50.126156Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-10T10:01:50.119265Z", "last_refresh": "2026-03-10T10:02:46.257769Z", "running": 8, "size": 8}}, {"events": ["2026-03-10T10:01:40.825549Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:00:52.912255Z", "last_refresh": "2026-03-10T10:02:47.314912Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:02:51.482921Z service:rgw.foorgw [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foorgw", "service_name": "rgw.foorgw", "service_type": "rgw", "spec": {"rgw_frontend_port": 8800}, "status": {"created": "2026-03-10T10:02:51.471546Z", "ports": [8800], "running": 0, "size": 2}}] 2026-03-10T10:03:17.277 INFO:tasks.cephadm:nfs.foo has 0/1 2026-03-10T10:03:17.785 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:17 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:17.786 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:17 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:17.786 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:17 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:17.786 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:17 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:17.786 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:17 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:03:18.091 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:17 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:18.091 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:17 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:18.091 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:17 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:18.091 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:17 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:18.091 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:17 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:03:18.278 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch ls -f json 2026-03-10T10:03:18.548 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:18.835 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:18 vm06.local ceph-mon[53281]: from='client.14674 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:18.835 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:18 vm06.local ceph-mon[53281]: pgmap v88: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 11 KiB/s rd, 709 B/s wr, 16 op/s 2026-03-10T10:03:18.835 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:18 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:18.835 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:18 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:18.840 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:03:18.840 INFO:teuthology.orchestra.run.vm06.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:00:54.272754Z", "last_refresh": "2026-03-10T10:02:47.314857Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:01:34.251508Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-10T10:00:52.371303Z", "last_refresh": "2026-03-10T10:02:47.314745Z", "ports": [9926], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:35.268370Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:00:51.930427Z", "last_refresh": "2026-03-10T10:02:47.314785Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-10T10:00:53.376036Z", "last_refresh": "2026-03-10T10:02:47.314884Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:03:16.570620Z service:ingress.nfs.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.foo", "service_name": "ingress.nfs.foo", "service_type": "ingress", "spec": {"backend_service": "nfs.foo", "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.106/22"}, "status": {"created": "2026-03-10T10:02:48.782609Z", "last_refresh": "2026-03-10T10:03:18.478882Z", "ports": [2049, 9049], "running": 2, "size": 4, "virtual_ip": "12.12.1.106/22"}}, {"events": ["2026-03-10T10:01:39.184549Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:00:51.537793Z", "last_refresh": "2026-03-10T10:02:47.314709Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:40.822470Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06:192.168.123.106=vm06", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:01:08.713712Z", "last_refresh": "2026-03-10T10:02:47.314648Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:02:52.951281Z service:nfs.foo [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "foo", "service_name": "nfs.foo", "service_type": "nfs", "spec": {"port": 12049}, "status": {"created": "2026-03-10T10:02:48.775768Z", "ports": [12049], "running": 0, "size": 1}}, {"events": ["2026-03-10T10:01:38.185414Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:00:53.882721Z", "last_refresh": "2026-03-10T10:02:47.314823Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:50.126156Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-10T10:01:50.119265Z", "last_refresh": "2026-03-10T10:02:47.314941Z", "running": 8, "size": 8}}, {"events": ["2026-03-10T10:01:40.825549Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:00:52.912255Z", "last_refresh": "2026-03-10T10:02:47.314912Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:02:51.482921Z service:rgw.foorgw [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foorgw", "service_name": "rgw.foorgw", "service_type": "rgw", "spec": {"rgw_frontend_port": 8800}, "status": {"created": "2026-03-10T10:02:51.471546Z", "last_refresh": "2026-03-10T10:03:18.478855Z", "ports": [8800], "running": 1, "size": 2}}] 2026-03-10T10:03:19.010 INFO:tasks.cephadm:nfs.foo has 0/1 2026-03-10T10:03:19.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:18 vm09.local ceph-mon[54558]: from='client.14674 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:19.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:18 vm09.local ceph-mon[54558]: pgmap v88: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 11 KiB/s rd, 709 B/s wr, 16 op/s 2026-03-10T10:03:19.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:18 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:19.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:18 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:20.011 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch ls -f json 2026-03-10T10:03:20.043 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:19 vm06.local ceph-mon[53281]: from='client.14678 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:20.043 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:19 vm06.local ceph-mon[53281]: pgmap v89: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 9.9 KiB/s rd, 170 B/s wr, 13 op/s 2026-03-10T10:03:20.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:19 vm09.local ceph-mon[54558]: from='client.14678 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:20.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:19 vm09.local ceph-mon[54558]: pgmap v89: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 9.9 KiB/s rd, 170 B/s wr, 13 op/s 2026-03-10T10:03:20.256 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:20.809 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:03:20.809 INFO:teuthology.orchestra.run.vm06.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:00:54.272754Z", "last_refresh": "2026-03-10T10:03:19.831852Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:01:34.251508Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-10T10:00:52.371303Z", "last_refresh": "2026-03-10T10:03:18.478551Z", "ports": [9926], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:35.268370Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:00:51.930427Z", "last_refresh": "2026-03-10T10:03:18.478615Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-10T10:00:53.376036Z", "last_refresh": "2026-03-10T10:03:19.831880Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:03:16.570620Z service:ingress.nfs.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.foo", "service_name": "ingress.nfs.foo", "service_type": "ingress", "spec": {"backend_service": "nfs.foo", "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.106/22"}, "status": {"created": "2026-03-10T10:02:48.782609Z", "last_refresh": "2026-03-10T10:03:18.478882Z", "ports": [2049, 9049], "running": 4, "size": 4, "virtual_ip": "12.12.1.106/22"}}, {"events": ["2026-03-10T10:01:39.184549Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:00:51.537793Z", "last_refresh": "2026-03-10T10:03:18.478680Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:40.822470Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06:192.168.123.106=vm06", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:01:08.713712Z", "last_refresh": "2026-03-10T10:03:18.478709Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:03:20.229073Z service:nfs.foo [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "foo", "service_name": "nfs.foo", "service_type": "nfs", "spec": {"port": 12049}, "status": {"created": "2026-03-10T10:02:48.775768Z", "last_refresh": "2026-03-10T10:03:19.832071Z", "ports": [12049], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:01:38.185414Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:00:53.882721Z", "last_refresh": "2026-03-10T10:03:18.478650Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:50.126156Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-10T10:01:50.119265Z", "last_refresh": "2026-03-10T10:03:18.478737Z", "running": 8, "size": 8}}, {"events": ["2026-03-10T10:01:40.825549Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:00:52.912255Z", "last_refresh": "2026-03-10T10:03:19.831907Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:02:51.482921Z service:rgw.foorgw [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foorgw", "service_name": "rgw.foorgw", "service_type": "rgw", "spec": {"rgw_frontend_port": 8800}, "status": {"created": "2026-03-10T10:02:51.471546Z", "last_refresh": "2026-03-10T10:03:18.478855Z", "ports": [8800], "running": 2, "size": 2}}] 2026-03-10T10:03:21.099 INFO:tasks.cephadm:nfs.foo has 1/1 2026-03-10T10:03:21.099 INFO:teuthology.run_tasks:Running task cephadm.wait_for_service... 2026-03-10T10:03:21.102 INFO:tasks.cephadm:Waiting for ceph service ingress.nfs.foo to start (timeout 300)... 2026-03-10T10:03:21.102 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph orch ls -f json 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: Checking dashboard <-> RGW credentials 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: Reconfiguring prometheus.vm06 (dependencies changed)... 2026-03-10T10:03:21.135 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:20 vm06.local ceph-mon[53281]: Reconfiguring daemon prometheus.vm06 on vm06 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: Checking dashboard <-> RGW credentials 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: Reconfiguring prometheus.vm06 (dependencies changed)... 2026-03-10T10:03:21.240 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:20 vm09.local ceph-mon[54558]: Reconfiguring daemon prometheus.vm06 on vm06 2026-03-10T10:03:21.560 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:21.851 INFO:teuthology.orchestra.run.vm06.stdout: 2026-03-10T10:03:21.851 INFO:teuthology.orchestra.run.vm06.stdout:[{"placement": {"count": 1}, "service_name": "alertmanager", "service_type": "alertmanager", "status": {"created": "2026-03-10T10:00:54.272754Z", "last_refresh": "2026-03-10T10:03:19.831852Z", "ports": [9093, 9094], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:01:34.251508Z service:ceph-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "ceph-exporter", "service_type": "ceph-exporter", "spec": {"prio_limit": 5, "stats_period": 5}, "status": {"created": "2026-03-10T10:00:52.371303Z", "last_refresh": "2026-03-10T10:03:18.478551Z", "ports": [9926], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:35.268370Z service:crash [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "crash", "service_type": "crash", "status": {"created": "2026-03-10T10:00:51.930427Z", "last_refresh": "2026-03-10T10:03:18.478615Z", "running": 2, "size": 2}}, {"placement": {"count": 1}, "service_name": "grafana", "service_type": "grafana", "spec": {"anonymous_access": true, "protocol": "https"}, "status": {"created": "2026-03-10T10:00:53.376036Z", "last_refresh": "2026-03-10T10:03:19.831880Z", "ports": [3000], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:03:16.570620Z service:ingress.nfs.foo [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "nfs.foo", "service_name": "ingress.nfs.foo", "service_type": "ingress", "spec": {"backend_service": "nfs.foo", "first_virtual_router_id": 50, "frontend_port": 2049, "monitor_port": 9049, "virtual_ip": "12.12.1.106/22"}, "status": {"created": "2026-03-10T10:02:48.782609Z", "last_refresh": "2026-03-10T10:03:18.478882Z", "ports": [2049, 9049], "running": 4, "size": 4, "virtual_ip": "12.12.1.106/22"}}, {"events": ["2026-03-10T10:01:39.184549Z service:mgr [INFO] \"service was created\""], "placement": {"count": 2}, "service_name": "mgr", "service_type": "mgr", "status": {"created": "2026-03-10T10:00:51.537793Z", "last_refresh": "2026-03-10T10:03:18.478680Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:40.822470Z service:mon [INFO] \"service was created\""], "placement": {"count": 2, "hosts": ["vm06:192.168.123.106=vm06", "vm09:192.168.123.109=vm09"]}, "service_name": "mon", "service_type": "mon", "status": {"created": "2026-03-10T10:01:08.713712Z", "last_refresh": "2026-03-10T10:03:18.478709Z", "running": 2, "size": 2}}, {"events": ["2026-03-10T10:03:20.229073Z service:nfs.foo [INFO] \"service was created\""], "placement": {"count": 1}, "service_id": "foo", "service_name": "nfs.foo", "service_type": "nfs", "spec": {"port": 12049}, "status": {"created": "2026-03-10T10:02:48.775768Z", "last_refresh": "2026-03-10T10:03:19.832071Z", "ports": [12049], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:01:38.185414Z service:node-exporter [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_name": "node-exporter", "service_type": "node-exporter", "status": {"created": "2026-03-10T10:00:53.882721Z", "last_refresh": "2026-03-10T10:03:18.478650Z", "ports": [9100], "running": 2, "size": 2}}, {"events": ["2026-03-10T10:01:50.126156Z service:osd.all-available-devices [INFO] \"service was created\""], "placement": {"host_pattern": "*"}, "service_id": "all-available-devices", "service_name": "osd.all-available-devices", "service_type": "osd", "spec": {"data_devices": {"all": true}, "filter_logic": "AND", "objectstore": "bluestore"}, "status": {"created": "2026-03-10T10:01:50.119265Z", "last_refresh": "2026-03-10T10:03:18.478737Z", "running": 8, "size": 8}}, {"events": ["2026-03-10T10:01:40.825549Z service:prometheus [INFO] \"service was created\""], "placement": {"count": 1}, "service_name": "prometheus", "service_type": "prometheus", "status": {"created": "2026-03-10T10:00:52.912255Z", "last_refresh": "2026-03-10T10:03:19.831907Z", "ports": [9095], "running": 1, "size": 1}}, {"events": ["2026-03-10T10:02:51.482921Z service:rgw.foorgw [INFO] \"service was created\""], "placement": {"count": 2}, "service_id": "foorgw", "service_name": "rgw.foorgw", "service_type": "rgw", "spec": {"rgw_frontend_port": 8800}, "status": {"created": "2026-03-10T10:02:51.471546Z", "last_refresh": "2026-03-10T10:03:18.478855Z", "ports": [8800], "running": 2, "size": 2}}] 2026-03-10T10:03:22.480 INFO:tasks.cephadm:ingress.nfs.foo has 4/4 2026-03-10T10:03:22.480 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-10T10:03:22.483 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm06.local 2026-03-10T10:03:22.483 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'mkdir /mnt/foo' 2026-03-10T10:03:22.511 INFO:teuthology.orchestra.run.vm06.stderr:+ mkdir /mnt/foo 2026-03-10T10:03:22.513 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'sleep 5' 2026-03-10T10:03:22.597 INFO:teuthology.orchestra.run.vm06.stderr:+ sleep 5 2026-03-10T10:03:22.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:22 vm09.local ceph-mon[54558]: from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:22.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:22 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:22.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:22 vm09.local ceph-mon[54558]: pgmap v90: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 5.6 KiB/s rd, 0 B/s wr, 7 op/s 2026-03-10T10:03:22.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:22 vm06.local ceph-mon[53281]: from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:22.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:22 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:22.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:22 vm06.local ceph-mon[53281]: pgmap v90: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 5.6 KiB/s rd, 0 B/s wr, 7 op/s 2026-03-10T10:03:23.725 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:23 vm09.local ceph-mon[54558]: from='client.14716 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:23.725 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:23 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:23.725 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:23 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:23.725 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:23 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:03:23.725 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:23 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:03:23.726 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:23 vm06.local ceph-mon[53281]: from='client.14716 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-10T10:03:23.726 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:23 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:23.726 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:23 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:23.726 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:23 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:03:23.726 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:23 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:03:24.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:24 vm09.local ceph-mon[54558]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:03:24.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:24 vm09.local ceph-mon[54558]: pgmap v91: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 5.5 KiB/s rd, 0 B/s wr, 6 op/s 2026-03-10T10:03:24.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:24 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:24.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:24 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:24.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:24 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:24.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:24 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:24.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:24 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:03:24.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:24 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:03:24.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:24 vm09.local ceph-mon[54558]: Checking dashboard <-> RGW credentials 2026-03-10T10:03:25.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:24 vm06.local ceph-mon[53281]: from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard get-prometheus-api-host"}]: dispatch 2026-03-10T10:03:25.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:24 vm06.local ceph-mon[53281]: pgmap v91: 225 pgs: 225 active+clean; 455 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 5.5 KiB/s rd, 0 B/s wr, 6 op/s 2026-03-10T10:03:25.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:24 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:25.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:24 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:25.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:24 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:25.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:24 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:25.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:24 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:03:25.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:24 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:03:25.068 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:24 vm06.local ceph-mon[53281]: Checking dashboard <-> RGW credentials 2026-03-10T10:03:26.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:26 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:26.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:26 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:03:26.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:26 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:26.318 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:26 vm06.local ceph-mon[53281]: pgmap v92: 225 pgs: 225 active+clean; 456 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-10T10:03:26.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:26 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:26.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:26 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:03:26.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:26 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:26.490 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:26 vm09.local ceph-mon[54558]: pgmap v92: 225 pgs: 225 active+clean; 456 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-10T10:03:27.598 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'mount -t nfs 12.12.1.106:/foobucket /mnt/foo' 2026-03-10T10:03:27.662 INFO:teuthology.orchestra.run.vm06.stderr:+ mount -t nfs 12.12.1.106:/foobucket /mnt/foo 2026-03-10T10:03:27.877 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'find /mnt/foo -ls' 2026-03-10T10:03:27.945 INFO:teuthology.orchestra.run.vm06.stderr:+ find /mnt/foo -ls 2026-03-10T10:03:27.951 INFO:teuthology.orchestra.run.vm06.stdout:18304478605914051910 0 drwxrwxrwx 1 root root 0 Mar 10 10:03 /mnt/foo 2026-03-10T10:03:27.952 INFO:teuthology.orchestra.run.vm06.stdout:10810806159419214356 0 -rw-rw-rw- 1 root root 7 Mar 10 10:03 /mnt/foo/myobject 2026-03-10T10:03:27.954 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'grep thebody /mnt/foo/myobject' 2026-03-10T10:03:28.021 INFO:teuthology.orchestra.run.vm06.stderr:+ grep thebody /mnt/foo/myobject 2026-03-10T10:03:28.024 INFO:teuthology.orchestra.run.vm06.stdout:thebody 2026-03-10T10:03:28.026 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'echo test > /mnt/foo/newobject' 2026-03-10T10:03:28.096 INFO:teuthology.orchestra.run.vm06.stderr:+ echo test 2026-03-10T10:03:28.300 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c sync 2026-03-10T10:03:28.368 INFO:teuthology.orchestra.run.vm06.stderr:+ sync 2026-03-10T10:03:28.378 INFO:teuthology.run_tasks:Running task python... 2026-03-10T10:03:28.381 INFO:tasks.python:Running python on role host.a host ubuntu@vm06.local 2026-03-10T10:03:28.381 INFO:tasks.python:import boto3 import json from io import BytesIO with open('/tmp/user.json', 'rt') as f: info = json.loads(f.read()) s3 = boto3.resource( 's3', aws_access_key_id=info['keys'][0]['access_key'], aws_secret_access_key=info['keys'][0]['secret_key'], endpoint_url='http://localhost:8800', ) bucket = s3.Bucket('foobucket') data = BytesIO() bucket.download_fileobj(Fileobj=data, Key='newobject') print(data.getvalue()) assert data.getvalue().decode() == 'test\n' 2026-03-10T10:03:28.381 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest python3 2026-03-10T10:03:28.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:28 vm06.local ceph-mon[53281]: pgmap v93: 225 pgs: 225 active+clean; 456 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 2.7 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-10T10:03:28.628 INFO:teuthology.orchestra.run.vm06.stdout:b'test\n' 2026-03-10T10:03:28.670 INFO:teuthology.run_tasks:Running task vip.exec... 2026-03-10T10:03:28.673 INFO:tasks.vip:Running commands on role host.a host ubuntu@vm06.local 2026-03-10T10:03:28.673 DEBUG:teuthology.orchestra.run.vm06:> sudo TESTDIR=/home/ubuntu/cephtest bash -ex -c 'umount /mnt/foo' 2026-03-10T10:03:28.699 INFO:teuthology.orchestra.run.vm06.stderr:+ umount /mnt/foo 2026-03-10T10:03:28.718 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T10:03:28.721 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm06.local 2026-03-10T10:03:28.721 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph nfs export rm foo /foobucket' 2026-03-10T10:03:28.739 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:28 vm09.local ceph-mon[54558]: pgmap v93: 225 pgs: 225 active+clean; 456 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 2.7 KiB/s rd, 170 B/s wr, 2 op/s 2026-03-10T10:03:28.953 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:29.423 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph nfs cluster rm foo' 2026-03-10T10:03:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:29 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.9", "id": [0, 2]}]: dispatch 2026-03-10T10:03:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:29 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 5]}]: dispatch 2026-03-10T10:03:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:29 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.17", "id": [0, 3]}]: dispatch 2026-03-10T10:03:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:29 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.c", "id": [3, 2]}]: dispatch 2026-03-10T10:03:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:29 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.18", "id": [7, 5]}]: dispatch 2026-03-10T10:03:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:29 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.e", "id": [1, 0]}]: dispatch 2026-03-10T10:03:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:29 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "5.5", "id": [4, 6]}]: dispatch 2026-03-10T10:03:29.568 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:29 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:03:29.621 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:29.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:29 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.9", "id": [0, 2]}]: dispatch 2026-03-10T10:03:29.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:29 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 5]}]: dispatch 2026-03-10T10:03:29.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:29 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.17", "id": [0, 3]}]: dispatch 2026-03-10T10:03:29.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:29 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.c", "id": [3, 2]}]: dispatch 2026-03-10T10:03:29.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:29 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.18", "id": [7, 5]}]: dispatch 2026-03-10T10:03:29.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:29 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.e", "id": [1, 0]}]: dispatch 2026-03-10T10:03:29.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:29 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "5.5", "id": [4, 6]}]: dispatch 2026-03-10T10:03:29.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:29 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd blocklist ls", "format": "json"}]: dispatch 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: pgmap v94: 225 pgs: 225 active+clean; 456 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 4.8 KiB/s rd, 170 B/s wr, 5 op/s 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: from='client.24455 -' entity='client.admin' cmd=[{"prefix": "nfs export rm", "cluster_id": "foo", "pseudo_path": "/foobucket", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.9", "id": [0, 2]}]': finished 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 5]}]': finished 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.17", "id": [0, 3]}]': finished 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.c", "id": [3, 2]}]': finished 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.18", "id": [7, 5]}]': finished 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.e", "id": [1, 0]}]': finished 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "5.5", "id": [4, 6]}]': finished 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: osdmap e47: 8 total, 8 up, 8 in 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:03:30.740 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:30 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:30.870 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-10T10:03:30.873 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm06.local 2026-03-10T10:03:30.873 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'stat -c '"'"'%u %g'"'"' /var/log/ceph | grep '"'"'167 167'"'"'' 2026-03-10T10:03:30.918 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: pgmap v94: 225 pgs: 225 active+clean; 456 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 4.8 KiB/s rd, 170 B/s wr, 5 op/s 2026-03-10T10:03:30.918 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: from='client.24455 -' entity='client.admin' cmd=[{"prefix": "nfs export rm", "cluster_id": "foo", "pseudo_path": "/foobucket", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:30.918 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.9", "id": [0, 2]}]': finished 2026-03-10T10:03:30.918 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.13", "id": [0, 5]}]': finished 2026-03-10T10:03:30.918 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "2.17", "id": [0, 3]}]': finished 2026-03-10T10:03:30.918 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.c", "id": [3, 2]}]': finished 2026-03-10T10:03:30.918 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "3.18", "id": [7, 5]}]': finished 2026-03-10T10:03:30.918 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "4.e", "id": [1, 0]}]': finished 2026-03-10T10:03:30.918 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd='[{"prefix": "osd pg-upmap-items", "format": "json", "pgid": "5.5", "id": [4, 6]}]': finished 2026-03-10T10:03:30.918 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: osdmap e47: 8 total, 8 up, 8 in 2026-03-10T10:03:30.919 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:30.919 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config dump", "format": "json"}]: dispatch 2026-03-10T10:03:30.919 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:30 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:31.195 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:31.306 INFO:teuthology.orchestra.run.vm06.stdout:167 167 2026-03-10T10:03:31.466 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch status' 2026-03-10T10:03:31.757 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:31.787 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:31 vm06.local ceph-mon[53281]: from='client.14746 -' entity='client.admin' cmd=[{"prefix": "nfs cluster rm", "cluster_id": "foo", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:31.787 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:31 vm06.local ceph-mon[53281]: Remove service ingress.nfs.foo 2026-03-10T10:03:31.787 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:31 vm06.local ceph-mon[53281]: Remove service nfs.foo 2026-03-10T10:03:31.787 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:31 vm06.local ceph-mon[53281]: osdmap e48: 8 total, 8 up, 8 in 2026-03-10T10:03:31.787 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:31 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:31.787 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:31 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:31.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:31 vm09.local ceph-mon[54558]: from='client.14746 -' entity='client.admin' cmd=[{"prefix": "nfs cluster rm", "cluster_id": "foo", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:31.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:31 vm09.local ceph-mon[54558]: Remove service ingress.nfs.foo 2026-03-10T10:03:31.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:31 vm09.local ceph-mon[54558]: Remove service nfs.foo 2026-03-10T10:03:31.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:31 vm09.local ceph-mon[54558]: osdmap e48: 8 total, 8 up, 8 in 2026-03-10T10:03:31.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:31 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:31.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:31 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:32.055 INFO:teuthology.orchestra.run.vm06.stdout:Backend: cephadm 2026-03-10T10:03:32.055 INFO:teuthology.orchestra.run.vm06.stdout:Available: Yes 2026-03-10T10:03:32.055 INFO:teuthology.orchestra.run.vm06.stdout:Paused: No 2026-03-10T10:03:32.269 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch ps' 2026-03-10T10:03:32.580 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:alertmanager.vm06 vm06 *:9093,9094 running (103s) 0s ago 2m 23.6M - 0.25.0 c8568f914cd2 2515739b1215 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:ceph-exporter.vm06 vm06 *:9926 running (2m) 0s ago 2m 9407k - 19.2.3-678-ge911bdeb 654f31e6858e a5751c00e778 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:ceph-exporter.vm09 vm09 *:9926 running (118s) 1s ago 118s 6614k - 19.2.3-678-ge911bdeb 654f31e6858e f34a9942447f 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:crash.vm06 vm06 running (2m) 0s ago 2m 7650k - 19.2.3-678-ge911bdeb 654f31e6858e 67ac1dfce5d4 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:crash.vm09 vm09 running (117s) 1s ago 117s 7658k - 19.2.3-678-ge911bdeb 654f31e6858e cb9dacbe6081 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:grafana.vm06 vm06 *:3000 running (101s) 0s ago 2m 70.9M - 10.4.0 c8b91775d855 be57c5781783 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:haproxy.nfs.foo.vm06.meexod vm06 *:2049,9049 running (35s) 0s ago 35s 3804k - 2.3.17-d1c9119 e85424b0d443 d6cc5ec311b0 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:haproxy.nfs.foo.vm09.aharkz vm09 *:2049,9049 running (30s) 1s ago 30s 3791k - 2.3.17-d1c9119 e85424b0d443 a7f82fe3522f 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:keepalived.nfs.foo.vm06.ulqfeo vm06 running (21s) 0s ago 21s 2391k - 2.2.4 4a3a1ff181d9 4f594da01df1 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:keepalived.nfs.foo.vm09.kzsmdy vm09 running (16s) 1s ago 16s 2391k - 2.2.4 4a3a1ff181d9 1e6bd030a915 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:mgr.vm06.dkjjvn vm06 *:9283,8765,8443 running (2m) 0s ago 2m 565M - 19.2.3-678-ge911bdeb 654f31e6858e 80e5643fbe08 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:mgr.vm09.jzyfcu vm09 *:8443,9283,8765 running (113s) 1s ago 113s 490M - 19.2.3-678-ge911bdeb 654f31e6858e 88a536ef8ea4 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:mon.vm06 vm06 running (2m) 0s ago 2m 55.7M 2048M 19.2.3-678-ge911bdeb 654f31e6858e f501e0d9f8a9 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:mon.vm09 vm09 running (112s) 1s ago 112s 49.2M 2048M 19.2.3-678-ge911bdeb 654f31e6858e 45d247861a5a 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:nfs.foo.0.0.vm06.giwwfx vm06 *:12049 running (40s) 0s ago 39s 117M - 5.9 654f31e6858e 434f2c87dd72 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:node-exporter.vm06 vm06 *:9100 running (2m) 0s ago 2m 9667k - 1.7.0 72c9c2088986 dde2f9d5f123 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:node-exporter.vm09 vm09 *:9100 running (114s) 1s ago 114s 9961k - 1.7.0 72c9c2088986 63d7ddb381cf 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:osd.0 vm09 running (77s) 1s ago 77s 49.9M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 09c6b7043ed8 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:osd.1 vm06 running (77s) 0s ago 77s 72.5M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 8125ae25527f 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:osd.2 vm06 running (74s) 0s ago 74s 50.7M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 394b47653578 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:osd.3 vm09 running (74s) 1s ago 74s 68.5M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 96ad4ab42962 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:osd.4 vm06 running (72s) 0s ago 72s 51.5M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 2469329ac060 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:osd.5 vm09 running (71s) 1s ago 71s 67.9M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 26c9ae68ff2f 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:osd.6 vm06 running (69s) 0s ago 68s 48.1M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 4748ed2672ec 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:osd.7 vm09 running (69s) 1s ago 68s 68.0M 4096M 19.2.3-678-ge911bdeb 654f31e6858e 7afa95a15e8f 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:prometheus.vm06 vm06 *:9095 running (10s) 0s ago 2m 32.7M - 2.51.0 1d3b7f56885b 936eff8ec817 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:rgw.foorgw.vm06.iyncag vm06 *:8800 running (41s) 0s ago 41s 102M - 19.2.3-678-ge911bdeb 654f31e6858e fff40f329caa 2026-03-10T10:03:32.876 INFO:teuthology.orchestra.run.vm06.stdout:rgw.foorgw.vm09.waheeg vm09 *:8800 running (42s) 1s ago 42s 99.2M - 19.2.3-678-ge911bdeb 654f31e6858e a02b4aa209ba 2026-03-10T10:03:32.877 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:32 vm06.local ceph-mon[53281]: pgmap v97: 225 pgs: 225 active+clean; 456 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 8.9 KiB/s rd, 767 B/s wr, 10 op/s 2026-03-10T10:03:32.877 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:32 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:32.877 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:32 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:32.877 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:32 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:03:32.877 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:32 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:03:32.877 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:32 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:32.877 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:32 vm06.local ceph-mon[53281]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:03:32.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:32 vm09.local ceph-mon[54558]: pgmap v97: 225 pgs: 225 active+clean; 456 KiB data, 223 MiB used, 160 GiB / 160 GiB avail; 8.9 KiB/s rd, 767 B/s wr, 10 op/s 2026-03-10T10:03:32.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:32 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:32.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:32 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:32.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:32 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "config generate-minimal-conf"}]: dispatch 2026-03-10T10:03:32.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:32 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "auth get", "entity": "client.admin"}]: dispatch 2026-03-10T10:03:32.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:32 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' 2026-03-10T10:03:32.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:32 vm09.local ceph-mon[54558]: from='mgr.14215 192.168.123.106:0/3070029319' entity='mgr.vm06.dkjjvn' cmd=[{"prefix": "osd tree", "states": ["destroyed"], "format": "json"}]: dispatch 2026-03-10T10:03:33.064 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch ls' 2026-03-10T10:03:33.252 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:33.522 INFO:teuthology.orchestra.run.vm06.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-10T10:03:33.522 INFO:teuthology.orchestra.run.vm06.stdout:alertmanager ?:9093,9094 1/1 0s ago 2m count:1 2026-03-10T10:03:33.522 INFO:teuthology.orchestra.run.vm06.stdout:ceph-exporter ?:9926 2/2 2s ago 2m * 2026-03-10T10:03:33.522 INFO:teuthology.orchestra.run.vm06.stdout:crash 2/2 2s ago 2m * 2026-03-10T10:03:33.522 INFO:teuthology.orchestra.run.vm06.stdout:grafana ?:3000 1/1 0s ago 2m count:1 2026-03-10T10:03:33.522 INFO:teuthology.orchestra.run.vm06.stdout:ingress.nfs.foo 12.12.1.106:2049,9049 4/4 44s count:2 2026-03-10T10:03:33.522 INFO:teuthology.orchestra.run.vm06.stdout:mgr 2/2 2s ago 2m count:2 2026-03-10T10:03:33.522 INFO:teuthology.orchestra.run.vm06.stdout:mon 2/2 2s ago 2m vm06:192.168.123.106=vm06;vm09:192.168.123.109=vm09;count:2 2026-03-10T10:03:33.522 INFO:teuthology.orchestra.run.vm06.stdout:nfs.foo ?:12049 1/1 44s count:1 2026-03-10T10:03:33.522 INFO:teuthology.orchestra.run.vm06.stdout:node-exporter ?:9100 2/2 2s ago 2m * 2026-03-10T10:03:33.522 INFO:teuthology.orchestra.run.vm06.stdout:osd.all-available-devices 8 2s ago 103s * 2026-03-10T10:03:33.523 INFO:teuthology.orchestra.run.vm06.stdout:prometheus ?:9095 1/1 0s ago 2m count:1 2026-03-10T10:03:33.523 INFO:teuthology.orchestra.run.vm06.stdout:rgw.foorgw ?:8800 2/2 2s ago 42s count:2 2026-03-10T10:03:33.704 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch host ls' 2026-03-10T10:03:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:33 vm06.local ceph-mon[53281]: from='client.14750 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:33 vm06.local ceph-mon[53281]: Removing orphan daemon nfs.foo.0.0.vm06.giwwfx... 2026-03-10T10:03:33.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:33 vm06.local ceph-mon[53281]: Removing daemon nfs.foo.0.0.vm06.giwwfx from vm06 -- ports [12049] 2026-03-10T10:03:33.898 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:33.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:33 vm09.local ceph-mon[54558]: from='client.14750 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:33.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:33 vm09.local ceph-mon[54558]: Removing orphan daemon nfs.foo.0.0.vm06.giwwfx... 2026-03-10T10:03:33.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:33 vm09.local ceph-mon[54558]: Removing daemon nfs.foo.0.0.vm06.giwwfx from vm06 -- ports [12049] 2026-03-10T10:03:34.148 INFO:teuthology.orchestra.run.vm06.stdout:HOST ADDR LABELS STATUS 2026-03-10T10:03:34.148 INFO:teuthology.orchestra.run.vm06.stdout:vm06 192.168.123.106 2026-03-10T10:03:34.148 INFO:teuthology.orchestra.run.vm06.stdout:vm09 192.168.123.109 2026-03-10T10:03:34.148 INFO:teuthology.orchestra.run.vm06.stdout:2 hosts in cluster 2026-03-10T10:03:34.314 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch device ls' 2026-03-10T10:03:34.502 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:34.750 INFO:teuthology.orchestra.run.vm06.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-10T10:03:34.750 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 45s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T10:03:34.750 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vdb hdd DWNBRSTVMM06001 20.0G No 45s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:03:34.750 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vdc hdd DWNBRSTVMM06002 20.0G No 45s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:03:34.750 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vdd hdd DWNBRSTVMM06003 20.0G No 45s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:03:34.750 INFO:teuthology.orchestra.run.vm06.stdout:vm06 /dev/vde hdd DWNBRSTVMM06004 20.0G No 45s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:03:34.750 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 46s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-10T10:03:34.751 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vdb hdd DWNBRSTVMM09001 20.0G No 46s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:03:34.751 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vdc hdd DWNBRSTVMM09002 20.0G No 46s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:03:34.751 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vdd hdd DWNBRSTVMM09003 20.0G No 46s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:03:34.751 INFO:teuthology.orchestra.run.vm06.stdout:vm09 /dev/vde hdd DWNBRSTVMM09004 20.0G No 46s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-10T10:03:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:34 vm06.local ceph-mon[53281]: from='client.14754 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:34 vm06.local ceph-mon[53281]: pgmap v98: 225 pgs: 225 active+clean; 456 KiB data, 224 MiB used, 160 GiB / 160 GiB avail; 14 KiB/s rd, 1023 B/s wr, 19 op/s; 13 B/s, 0 objects/s recovering 2026-03-10T10:03:34.818 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:34 vm06.local ceph-mon[53281]: from='client.14758 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:34.914 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- bash -c 'ceph orch ls | grep '"'"'^osd.all-available-devices '"'"'' 2026-03-10T10:03:34.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:34 vm09.local ceph-mon[54558]: from='client.14754 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:34.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:34 vm09.local ceph-mon[54558]: pgmap v98: 225 pgs: 225 active+clean; 456 KiB data, 224 MiB used, 160 GiB / 160 GiB avail; 14 KiB/s rd, 1023 B/s wr, 19 op/s; 13 B/s, 0 objects/s recovering 2026-03-10T10:03:34.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:34 vm09.local ceph-mon[54558]: from='client.14758 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:35.115 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:35.396 INFO:teuthology.orchestra.run.vm06.stdout:osd.all-available-devices 8 3s ago 105s * 2026-03-10T10:03:35.574 DEBUG:teuthology.run_tasks:Unwinding manager vip 2026-03-10T10:03:35.577 INFO:tasks.vip:Removing 12.12.0.106 (and any VIPs) on vm06.local iface eth0... 2026-03-10T10:03:35.577 DEBUG:teuthology.orchestra.run.vm06:> sudo ip addr del 12.12.0.106/22 dev eth0 2026-03-10T10:03:35.602 DEBUG:teuthology.orchestra.run.vm06:> sudo ip addr del 12.12.1.106/22 dev eth0 2026-03-10T10:03:35.673 INFO:tasks.vip:Removing 12.12.0.109 (and any VIPs) on vm09.local iface eth0... 2026-03-10T10:03:35.673 DEBUG:teuthology.orchestra.run.vm09:> sudo ip addr del 12.12.0.109/22 dev eth0 2026-03-10T10:03:35.709 DEBUG:teuthology.orchestra.run.vm09:> sudo ip addr del 12.12.1.106/22 dev eth0 2026-03-10T10:03:35.775 INFO:teuthology.orchestra.run.vm09.stderr:Error: ipv4: Address not found. 2026-03-10T10:03:35.776 DEBUG:teuthology.orchestra.run:got remote process result: 2 2026-03-10T10:03:35.776 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-10T10:03:35.779 INFO:tasks.cephadm:Teardown begin 2026-03-10T10:03:35.779 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:03:35.811 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:03:35.843 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-10T10:03:35.843 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 -- ceph mgr module disable cephadm 2026-03-10T10:03:35.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:35 vm09.local ceph-mon[54558]: from='client.14762 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:36.048 INFO:teuthology.orchestra.run.vm06.stderr:Inferring config /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/mon.vm06/config 2026-03-10T10:03:36.058 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:35 vm06.local ceph-mon[53281]: from='client.14762 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-10T10:03:36.096 INFO:teuthology.orchestra.run.vm06.stderr:Error: statfs /etc/ceph/ceph.client.admin.keyring: no such file or directory 2026-03-10T10:03:36.120 DEBUG:teuthology.orchestra.run:got remote process result: 125 2026-03-10T10:03:36.120 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-10T10:03:36.120 DEBUG:teuthology.orchestra.run.vm06:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T10:03:36.136 DEBUG:teuthology.orchestra.run.vm09:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-10T10:03:36.152 INFO:tasks.cephadm:Stopping all daemons... 2026-03-10T10:03:36.152 INFO:tasks.cephadm.mon.vm06:Stopping mon.vm06... 2026-03-10T10:03:36.152 DEBUG:teuthology.orchestra.run.vm06:> sudo systemctl stop ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mon.vm06 2026-03-10T10:03:36.310 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:36 vm06.local systemd[1]: Stopping Ceph mon.vm06 for ddb3da90-1c67-11f1-bef5-37be0d23c456... 2026-03-10T10:03:36.566 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:36 vm06.local ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456-mon-vm06[53256]: 2026-03-10T10:03:36.304+0000 7fc00ae5e640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.vm06 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T10:03:36.566 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:36 vm06.local ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456-mon-vm06[53256]: 2026-03-10T10:03:36.304+0000 7fc00ae5e640 -1 mon.vm06@0(leader) e2 *** Got Signal Terminated *** 2026-03-10T10:03:36.567 INFO:journalctl@ceph.mon.vm06.vm06.stdout:Mar 10 10:03:36 vm06.local podman[98514]: 2026-03-10 10:03:36.430674188 +0000 UTC m=+0.143120344 container died f501e0d9f8a9ca1d3bf0571e267ccc06d2722c11ec96ee31dd8865a3e4a35679 (image=quay.ceph.io/ceph-ci/ceph:e911bdebe5c8faa3800735d1568fcdca65db60df, name=ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456-mon-vm06, org.label-schema.vendor=CentOS, OSD_FLAVOR=default, ceph=True, org.opencontainers.image.documentation=https://docs.ceph.com/, org.label-schema.build-date=20260223, org.label-schema.license=GPLv2, CEPH_GIT_REPO=https://github.com/ceph/ceph-ci.git, FROM_IMAGE=quay.io/centos/centos:stream9, org.label-schema.name=CentOS Stream 9 Base Image, org.label-schema.schema-version=1.0, GANESHA_REPO_BASEURL=https://buildlogs.centos.org/centos/$releasever-stream/storage/$basearch/nfsganesha-5/, io.buildah.version=1.41.3, org.opencontainers.image.authors=Ceph Release Team , CEPH_REF=squid, CEPH_SHA1=e911bdebe5c8faa3800735d1568fcdca65db60df) 2026-03-10T10:03:36.659 DEBUG:teuthology.orchestra.run.vm06:> sudo pkill -f 'journalctl -f -n 0 -u ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mon.vm06.service' 2026-03-10T10:03:36.704 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T10:03:36.704 INFO:tasks.cephadm.mon.vm06:Stopped mon.vm06 2026-03-10T10:03:36.704 INFO:tasks.cephadm.mon.vm09:Stopping mon.vm09... 2026-03-10T10:03:36.704 DEBUG:teuthology.orchestra.run.vm09:> sudo systemctl stop ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mon.vm09 2026-03-10T10:03:36.990 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:36 vm09.local systemd[1]: Stopping Ceph mon.vm09 for ddb3da90-1c67-11f1-bef5-37be0d23c456... 2026-03-10T10:03:36.991 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:36 vm09.local ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456-mon-vm09[54534]: 2026-03-10T10:03:36.851+0000 7f9fa8c39640 -1 received signal: Terminated from /run/podman-init -- /usr/bin/ceph-mon -n mon.vm09 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-journald=true --default-log-to-stderr=false --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-journald=true --default-mon-cluster-log-to-stderr=false (PID: 1) UID: 0 2026-03-10T10:03:36.991 INFO:journalctl@ceph.mon.vm09.vm09.stdout:Mar 10 10:03:36 vm09.local ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456-mon-vm09[54534]: 2026-03-10T10:03:36.851+0000 7f9fa8c39640 -1 mon.vm09@1(peon) e2 *** Got Signal Terminated *** 2026-03-10T10:03:37.270 DEBUG:teuthology.orchestra.run.vm09:> sudo pkill -f 'journalctl -f -n 0 -u ceph-ddb3da90-1c67-11f1-bef5-37be0d23c456@mon.vm09.service' 2026-03-10T10:03:37.311 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-10T10:03:37.311 INFO:tasks.cephadm.mon.vm09:Stopped mon.vm09 2026-03-10T10:03:37.311 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 --force --keep-logs 2026-03-10T10:03:43.386 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:04:23.794 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 --force --keep-logs 2026-03-10T10:04:23.939 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:05:03.462 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:05:03.487 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-10T10:05:03.512 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-10T10:05:03.513 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/992/remote/vm06/crash 2026-03-10T10:05:03.513 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/crash -- . 2026-03-10T10:05:03.550 INFO:teuthology.orchestra.run.vm06.stderr:tar: /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/crash: Cannot open: No such file or directory 2026-03-10T10:05:03.550 INFO:teuthology.orchestra.run.vm06.stderr:tar: Error is not recoverable: exiting now 2026-03-10T10:05:03.551 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/crash to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/992/remote/vm09/crash 2026-03-10T10:05:03.551 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/crash -- . 2026-03-10T10:05:03.577 INFO:teuthology.orchestra.run.vm09.stderr:tar: /var/lib/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/crash: Cannot open: No such file or directory 2026-03-10T10:05:03.577 INFO:teuthology.orchestra.run.vm09.stderr:tar: Error is not recoverable: exiting now 2026-03-10T10:05:03.578 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-10T10:05:03.578 DEBUG:teuthology.orchestra.run.vm06:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.log | egrep CEPHADM_ | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v CEPHADM_DAEMON_PLACE_FAIL | egrep -v CEPHADM_FAILED_DAEMON | head -n 1 2026-03-10T10:05:03.617 INFO:tasks.cephadm:Compressing logs... 2026-03-10T10:05:03.618 DEBUG:teuthology.orchestra.run.vm06:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T10:05:03.660 DEBUG:teuthology.orchestra.run.vm09:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T10:05:03.683 INFO:teuthology.orchestra.run.vm09.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T10:05:03.683 INFO:teuthology.orchestra.run.vm09.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T10:05:03.683 INFO:teuthology.orchestra.run.vm06.stderr:find: gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-10T10:05:03.683 INFO:teuthology.orchestra.run.vm06.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-10T10:05:03.684 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-volume.log 2026-03-10T10:05:03.684 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.ceph-exporter.vm09.log 2026-03-10T10:05:03.684 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-volume.log: 91.6% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T10:05:03.684 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/cephadm.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mon.vm06.log 2026-03-10T10:05:03.685 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mgr.vm09.jzyfcu.log 2026-03-10T10:05:03.685 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.ceph-exporter.vm09.log: 28.6% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.ceph-exporter.vm09.log.gz 2026-03-10T10:05:03.689 INFO:teuthology.orchestra.run.vm09.stderr: 95.7% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-volume.log.gz 2026-03-10T10:05:03.689 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mon.vm09.log 2026-03-10T10:05:03.689 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mgr.vm09.jzyfcu.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.audit.log 2026-03-10T10:05:03.690 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mon.vm06.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.log 2026-03-10T10:05:03.691 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mon.vm09.log: 91.3% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mgr.vm09.jzyfcu.log.gz 2026-03-10T10:05:03.691 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.log 2026-03-10T10:05:03.692 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.audit.log: 91.0% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.audit.log.gz 2026-03-10T10:05:03.693 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.cephadm.log 2026-03-10T10:05:03.693 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.log: 83.3% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.log.gz 2026-03-10T10:05:03.693 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.0.log 2026-03-10T10:05:03.694 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.cephadm.log: 81.7% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.cephadm.log.gz 2026-03-10T10:05:03.694 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.3.log 2026-03-10T10:05:03.697 INFO:teuthology.orchestra.run.vm06.stderr: 91.5% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-10T10:05:03.697 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.audit.log 2026-03-10T10:05:03.697 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mgr.vm06.dkjjvn.log 2026-03-10T10:05:03.697 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.log: /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.audit.log: 84.2% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.log.gz 2026-03-10T10:05:03.699 INFO:teuthology.orchestra.run.vm06.stderr: 90.8% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.audit.log.gz 2026-03-10T10:05:03.699 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.cephadm.log 2026-03-10T10:05:03.699 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.0.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.5.log 2026-03-10T10:05:03.708 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mgr.vm06.dkjjvn.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-volume.log 2026-03-10T10:05:03.708 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.cephadm.log: 82.7% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph.cephadm.log.gz 2026-03-10T10:05:03.710 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.3.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.7.log 2026-03-10T10:05:03.713 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.ceph-exporter.vm06.log 2026-03-10T10:05:03.717 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.5.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.rgw.foorgw.vm09.waheeg.log 2026-03-10T10:05:03.720 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-volume.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.1.log 2026-03-10T10:05:03.720 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.ceph-exporter.vm06.log: 92.5% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.ceph-exporter.vm06.log.gz 2026-03-10T10:05:03.723 INFO:teuthology.orchestra.run.vm06.stderr: 95.7% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-volume.log.gz 2026-03-10T10:05:03.724 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.2.log 2026-03-10T10:05:03.727 INFO:teuthology.orchestra.run.vm09.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.7.log: /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.rgw.foorgw.vm09.waheeg.log: 58.6% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.rgw.foorgw.vm09.waheeg.log.gz 2026-03-10T10:05:03.729 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.1.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.4.log 2026-03-10T10:05:03.740 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.2.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.6.log 2026-03-10T10:05:03.744 INFO:teuthology.orchestra.run.vm09.stderr: 92.2% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mon.vm09.log.gz 2026-03-10T10:05:03.745 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.4.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.rgw.foorgw.vm06.iyncag.log 2026-03-10T10:05:03.756 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.6.log: gzip -5 --verbose -- /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ops-log-ceph-client.rgw.foorgw.vm06.iyncag.log 2026-03-10T10:05:03.758 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.rgw.foorgw.vm06.iyncag.log: 70.1% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-client.rgw.foorgw.vm06.iyncag.log.gz 2026-03-10T10:05:03.768 INFO:teuthology.orchestra.run.vm06.stderr:/var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ops-log-ceph-client.rgw.foorgw.vm06.iyncag.log: 77.7% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ops-log-ceph-client.rgw.foorgw.vm06.iyncag.log.gz 2026-03-10T10:05:03.838 INFO:teuthology.orchestra.run.vm06.stderr: 89.6% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mgr.vm06.dkjjvn.log.gz 2026-03-10T10:05:03.961 INFO:teuthology.orchestra.run.vm09.stderr: 93.9% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.0.log.gz 2026-03-10T10:05:03.976 INFO:teuthology.orchestra.run.vm06.stderr: 91.6% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-mon.vm06.log.gz 2026-03-10T10:05:03.997 INFO:teuthology.orchestra.run.vm09.stderr: 93.8% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.5.log.gz 2026-03-10T10:05:03.998 INFO:teuthology.orchestra.run.vm09.stderr: 93.8% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.7.log.gz 2026-03-10T10:05:04.020 INFO:teuthology.orchestra.run.vm09.stderr: 93.9% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.3.log.gz 2026-03-10T10:05:04.021 INFO:teuthology.orchestra.run.vm09.stderr: 2026-03-10T10:05:04.021 INFO:teuthology.orchestra.run.vm09.stderr:real 0m0.348s 2026-03-10T10:05:04.021 INFO:teuthology.orchestra.run.vm09.stderr:user 0m0.621s 2026-03-10T10:05:04.021 INFO:teuthology.orchestra.run.vm09.stderr:sys 0m0.040s 2026-03-10T10:05:04.026 INFO:teuthology.orchestra.run.vm06.stderr: 93.9% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.2.log.gz 2026-03-10T10:05:04.049 INFO:teuthology.orchestra.run.vm06.stderr: 94.2% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.4.log.gz 2026-03-10T10:05:04.057 INFO:teuthology.orchestra.run.vm06.stderr: 94.1% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.1.log.gz 2026-03-10T10:05:04.071 INFO:teuthology.orchestra.run.vm06.stderr: 93.9% -- replaced with /var/log/ceph/ddb3da90-1c67-11f1-bef5-37be0d23c456/ceph-osd.6.log.gz 2026-03-10T10:05:04.072 INFO:teuthology.orchestra.run.vm06.stderr: 2026-03-10T10:05:04.072 INFO:teuthology.orchestra.run.vm06.stderr:real 0m0.399s 2026-03-10T10:05:04.072 INFO:teuthology.orchestra.run.vm06.stderr:user 0m0.721s 2026-03-10T10:05:04.072 INFO:teuthology.orchestra.run.vm06.stderr:sys 0m0.053s 2026-03-10T10:05:04.073 INFO:tasks.cephadm:Archiving logs... 2026-03-10T10:05:04.073 DEBUG:teuthology.misc:Transferring archived files from vm06:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/992/remote/vm06/log 2026-03-10T10:05:04.073 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T10:05:04.183 DEBUG:teuthology.misc:Transferring archived files from vm09:/var/log/ceph to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/992/remote/vm09/log 2026-03-10T10:05:04.183 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-10T10:05:04.246 INFO:tasks.cephadm:Removing cluster... 2026-03-10T10:05:04.246 DEBUG:teuthology.orchestra.run.vm06:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 --force 2026-03-10T10:05:04.372 INFO:teuthology.orchestra.run.vm06.stdout:Deleting cluster with fsid: ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:05:04.472 DEBUG:teuthology.orchestra.run.vm09:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid ddb3da90-1c67-11f1-bef5-37be0d23c456 --force 2026-03-10T10:05:04.602 INFO:teuthology.orchestra.run.vm09.stdout:Deleting cluster with fsid: ddb3da90-1c67-11f1-bef5-37be0d23c456 2026-03-10T10:05:04.703 INFO:tasks.cephadm:Removing cephadm ... 2026-03-10T10:05:04.703 DEBUG:teuthology.orchestra.run.vm06:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T10:05:04.720 DEBUG:teuthology.orchestra.run.vm09:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-10T10:05:04.739 INFO:tasks.cephadm:Teardown complete 2026-03-10T10:05:04.739 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-10T10:05:04.742 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-10T10:05:04.742 DEBUG:teuthology.orchestra.run.vm06:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T10:05:04.762 DEBUG:teuthology.orchestra.run.vm09:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-10T10:05:04.777 INFO:teuthology.orchestra.run.vm06.stderr:bash: line 1: ntpq: command not found 2026-03-10T10:05:04.781 INFO:teuthology.orchestra.run.vm06.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T10:05:04.781 INFO:teuthology.orchestra.run.vm06.stdout:=============================================================================== 2026-03-10T10:05:04.781 INFO:teuthology.orchestra.run.vm06.stdout:^+ sv5.ggsrv.de 2 6 377 58 +2626us[+2626us] +/- 23ms 2026-03-10T10:05:04.781 INFO:teuthology.orchestra.run.vm06.stdout:^* netcup01.theravenhub.com 2 6 377 58 -1114us[-1229us] +/- 18ms 2026-03-10T10:05:04.781 INFO:teuthology.orchestra.run.vm06.stdout:^+ srv01-nc.securepod.org 2 6 377 56 -983us[ -983us] +/- 18ms 2026-03-10T10:05:04.781 INFO:teuthology.orchestra.run.vm06.stdout:^- mail.vbrandl.net 3 6 377 58 -2961us[-3075us] +/- 64ms 2026-03-10T10:05:04.796 INFO:teuthology.orchestra.run.vm09.stderr:bash: line 1: ntpq: command not found 2026-03-10T10:05:04.800 INFO:teuthology.orchestra.run.vm09.stdout:MS Name/IP address Stratum Poll Reach LastRx Last sample 2026-03-10T10:05:04.800 INFO:teuthology.orchestra.run.vm09.stdout:=============================================================================== 2026-03-10T10:05:04.800 INFO:teuthology.orchestra.run.vm09.stdout:^+ netcup01.theravenhub.com 2 6 377 60 -1128us[-1043us] +/- 18ms 2026-03-10T10:05:04.800 INFO:teuthology.orchestra.run.vm09.stdout:^* srv01-nc.securepod.org 2 6 377 57 -998us[ -913us] +/- 18ms 2026-03-10T10:05:04.800 INFO:teuthology.orchestra.run.vm09.stdout:^- mail.vbrandl.net 3 6 377 58 -2975us[-2889us] +/- 64ms 2026-03-10T10:05:04.800 INFO:teuthology.orchestra.run.vm09.stdout:^+ sv5.ggsrv.de 2 6 377 57 +2811us[+2811us] +/- 23ms 2026-03-10T10:05:04.800 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-10T10:05:04.803 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-10T10:05:04.803 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-10T10:05:04.806 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-10T10:05:04.808 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-10T10:05:04.811 INFO:teuthology.task.internal:Duration was 481.505417 seconds 2026-03-10T10:05:04.811 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-10T10:05:04.814 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-10T10:05:04.814 DEBUG:teuthology.orchestra.run.vm06:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T10:05:04.824 DEBUG:teuthology.orchestra.run.vm09:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-10T10:05:04.864 INFO:teuthology.orchestra.run.vm06.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T10:05:04.881 INFO:teuthology.orchestra.run.vm09.stderr:Redirecting to /bin/systemctl restart rsyslog.service 2026-03-10T10:05:05.342 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-10T10:05:05.342 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm06.local 2026-03-10T10:05:05.342 DEBUG:teuthology.orchestra.run.vm06:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T10:05:05.406 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm09.local 2026-03-10T10:05:05.407 DEBUG:teuthology.orchestra.run.vm09:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-10T10:05:05.437 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-10T10:05:05.437 DEBUG:teuthology.orchestra.run.vm06:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T10:05:05.449 DEBUG:teuthology.orchestra.run.vm09:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T10:05:06.016 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-10T10:05:06.016 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T10:05:06.017 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-10T10:05:06.041 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T10:05:06.041 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T10:05:06.041 INFO:teuthology.orchestra.run.vm06.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T10:05:06.041 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T10:05:06.041 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T10:05:06.043 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-10T10:05:06.043 INFO:teuthology.orchestra.run.vm09.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-10T10:05:06.044 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-10T10:05:06.044 INFO:teuthology.orchestra.run.vm09.stderr: -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-10T10:05:06.044 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-10T10:05:06.166 INFO:teuthology.orchestra.run.vm09.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T10:05:06.218 INFO:teuthology.orchestra.run.vm06.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 98.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-10T10:05:06.220 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-10T10:05:06.224 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-10T10:05:06.224 DEBUG:teuthology.orchestra.run.vm06:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T10:05:06.286 DEBUG:teuthology.orchestra.run.vm09:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-10T10:05:06.310 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-10T10:05:06.313 DEBUG:teuthology.orchestra.run.vm06:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T10:05:06.328 DEBUG:teuthology.orchestra.run.vm09:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-10T10:05:06.352 INFO:teuthology.orchestra.run.vm06.stdout:kernel.core_pattern = core 2026-03-10T10:05:06.378 INFO:teuthology.orchestra.run.vm09.stdout:kernel.core_pattern = core 2026-03-10T10:05:06.391 DEBUG:teuthology.orchestra.run.vm06:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T10:05:06.419 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:05:06.419 DEBUG:teuthology.orchestra.run.vm09:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-10T10:05:06.444 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-10T10:05:06.444 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-10T10:05:06.448 INFO:teuthology.task.internal:Transferring archived files... 2026-03-10T10:05:06.448 DEBUG:teuthology.misc:Transferring archived files from vm06:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/992/remote/vm06 2026-03-10T10:05:06.448 DEBUG:teuthology.orchestra.run.vm06:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T10:05:06.489 DEBUG:teuthology.misc:Transferring archived files from vm09:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-10_01:00:38-orch-squid-none-default-vps/992/remote/vm09 2026-03-10T10:05:06.489 DEBUG:teuthology.orchestra.run.vm09:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-10T10:05:06.516 INFO:teuthology.task.internal:Removing archive directory... 2026-03-10T10:05:06.516 DEBUG:teuthology.orchestra.run.vm06:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T10:05:06.530 DEBUG:teuthology.orchestra.run.vm09:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-10T10:05:06.571 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-10T10:05:06.574 INFO:teuthology.task.internal:Not uploading archives. 2026-03-10T10:05:06.574 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-10T10:05:06.578 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-10T10:05:06.578 DEBUG:teuthology.orchestra.run.vm06:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T10:05:06.586 DEBUG:teuthology.orchestra.run.vm09:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-10T10:05:06.600 INFO:teuthology.orchestra.run.vm06.stdout: 8532144 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 10:05 /home/ubuntu/cephtest 2026-03-10T10:05:06.627 INFO:teuthology.orchestra.run.vm09.stdout: 8532147 0 drwxr-xr-x 2 ubuntu ubuntu 6 Mar 10 10:05 /home/ubuntu/cephtest 2026-03-10T10:05:06.628 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-10T10:05:06.635 INFO:teuthology.run:Summary data: description: orch/cephadm/smoke-roleless/{0-distro/centos_9.stream_runc 1-start 2-services/nfs-ingress-rgw-bucket 3-final} duration: 481.5054168701172 owner: kyr success: true 2026-03-10T10:05:06.635 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-10T10:05:06.654 INFO:teuthology.run:pass