2026-03-21T06:44:01.604 INFO:root:teuthology version: 1.2.4.dev6+g1c580df7a 2026-03-21T06:44:01.609 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-21T06:44:01.627 INFO:teuthology.run:Config: archive_path: /archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458 branch: tentacle description: rbd/iscsi/{base/install cluster/{fixed-3} conf/{disable-pool-app} supported-container-hosts$/{ubuntu_22.04} workloads/cephadm_iscsi} email: null first_in_suite: false flavor: default job_id: '3458' ktype: distro last_in_suite: false machine_type: vps name: kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps no_nested_subset: false os_type: ubuntu os_version: '22.04' overrides: admin_socket: branch: tentacle ansible.cephlab: branch: main repo: https://github.com/kshtsk/ceph-cm-ansible.git skip_tags: nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs vars: logical_volumes: lv_1: scratch_dev: true size: 25%VG vg: vg_nvme lv_2: scratch_dev: true size: 25%VG vg: vg_nvme lv_3: scratch_dev: true size: 25%VG vg: vg_nvme lv_4: scratch_dev: true size: 25%VG vg: vg_nvme timezone: UTC volume_groups: vg_nvme: pvs: /dev/vdb,/dev/vdc,/dev/vdd,/dev/vde ceph: conf: global: mon warn on pool no app: false mgr: debug mgr: 20 debug ms: 1 mon: debug mon: 20 debug ms: 1 debug paxos: 20 osd: debug ms: 1 debug osd: 20 osd mclock iops capacity threshold hdd: 49000 flavor: default log-ignorelist: - \(MDS_ALL_DOWN\) - \(MDS_UP_LESS_THAN_MAX\) - MON_DOWN sha1: 70f8415b300f041766fa27faf7d5472699e32388 ceph-deploy: conf: client: log file: /var/log/ceph/ceph-$name.$pid.log mon: {} cephadm: cephadm_binary_url: https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm install: ceph: flavor: default sha1: 70f8415b300f041766fa27faf7d5472699e32388 extra_system_packages: deb: - python3-jmespath - python3-xmltodict - s3cmd rpm: - bzip2 - perl-Test-Harness - python3-jmespath - python3-xmltodict - s3cmd workunit: branch: tt-tentacle sha1: 0392f78529848ec72469e8e431875cb98d3a5fb4 owner: kyr priority: 1000 repo: https://github.com/ceph/ceph.git roles: - - host.a - mon.a - mgr.x - osd.0 - osd.1 - client.0 - ceph.iscsi.iscsi.a - - mon.b - osd.2 - osd.3 - osd.4 - client.1 - - mon.c - osd.5 - osd.6 - osd.7 - client.2 - ceph.iscsi.iscsi.b seed: 3051 sha1: 70f8415b300f041766fa27faf7d5472699e32388 sleep_before_teardown: 0 subset: 1/128 suite: rbd suite_branch: tt-tentacle suite_path: /home/teuthos/src/github.com_kshtsk_ceph_0392f78529848ec72469e8e431875cb98d3a5fb4/qa suite_relpath: qa suite_repo: https://github.com/kshtsk/ceph.git suite_sha1: 0392f78529848ec72469e8e431875cb98d3a5fb4 targets: vm02.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLcZNeO8u6CEcqUxixY+8HC17tcUZvEGRDk5nT+gHnfF8RoNMc0vTT8W+3sP5EKFyqmgPI7kcILj/IibNmZNa7c= vm04.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBD2x50LP3vGTd+y7mFC05mi/JRpXP/fcTXHJPS4mJXsdem9ZKd8/nGs2eRti3eHTdUiHgbNHam+5hy3IwloQqaY= vm07.local: ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMQCNACJyBC9mJ7lgcIhNI5gd78lRhE4sNx6XNqeQSGilD+wEpJHdUiikPQDbhYkQs5jIcdpYmXGauo70zoDR/k= tasks: - cephadm: null - cephadm.shell: host.a: - ceph orch status - ceph orch ps - ceph orch ls - ceph orch host ls - ceph orch device ls - install: extra_system_packages: deb: - open-iscsi - multipath-tools rpm: - iscsi-initiator-utils - device-mapper-multipath - ceph_iscsi_client: clients: - client.1 - cram: clients: client.0: - src/test/cli-integration/rbd/gwcli_create.t client.1: - src/test/cli-integration/rbd/iscsi_client.t client.2: - src/test/cli-integration/rbd/gwcli_delete.t parallel: false - cram: clients: client.0: - src/test/cli-integration/rbd/rest_api_create.t client.1: - src/test/cli-integration/rbd/iscsi_client.t client.2: - src/test/cli-integration/rbd/rest_api_delete.t parallel: false teuthology: fragments_dropped: [] meta: {} postmerge: [] teuthology_branch: clyso-debian-13 teuthology_repo: https://github.com/clyso/teuthology teuthology_sha1: 1c580df7a9c7c2aadc272da296344fd99f27c444 timestamp: 2026-03-20_22:04:26 tube: vps use_shaman: true user: kyr verbose: false worker_log: /home/teuthos/.teuthology/dispatcher/dispatcher.vps.4188345 2026-03-21T06:44:01.627 INFO:teuthology.run:suite_path is set to /home/teuthos/src/github.com_kshtsk_ceph_0392f78529848ec72469e8e431875cb98d3a5fb4/qa; will attempt to use it 2026-03-21T06:44:01.627 INFO:teuthology.run:Found tasks at /home/teuthos/src/github.com_kshtsk_ceph_0392f78529848ec72469e8e431875cb98d3a5fb4/qa/tasks 2026-03-21T06:44:01.627 INFO:teuthology.run_tasks:Running task internal.check_packages... 2026-03-21T06:44:01.628 INFO:teuthology.task.internal:Checking packages... 2026-03-21T06:44:01.628 INFO:teuthology.task.internal:Checking packages for os_type 'ubuntu', flavor 'default' and ceph hash '70f8415b300f041766fa27faf7d5472699e32388' 2026-03-21T06:44:01.628 WARNING:teuthology.packaging:More than one of ref, tag, branch, or sha1 supplied; using branch 2026-03-21T06:44:01.628 INFO:teuthology.packaging:ref: None 2026-03-21T06:44:01.628 INFO:teuthology.packaging:tag: None 2026-03-21T06:44:01.628 INFO:teuthology.packaging:branch: tentacle 2026-03-21T06:44:01.628 INFO:teuthology.packaging:sha1: 70f8415b300f041766fa27faf7d5472699e32388 2026-03-21T06:44:01.628 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&ref=tentacle 2026-03-21T06:44:02.408 INFO:teuthology.task.internal:Found packages for ceph version 20.2.0-714-g147f7c6a-1jammy 2026-03-21T06:44:02.409 INFO:teuthology.run_tasks:Running task internal.buildpackages_prep... 2026-03-21T06:44:02.410 INFO:teuthology.task.internal:no buildpackages task found 2026-03-21T06:44:02.410 INFO:teuthology.run_tasks:Running task internal.save_config... 2026-03-21T06:44:02.410 INFO:teuthology.task.internal:Saving configuration 2026-03-21T06:44:02.416 INFO:teuthology.run_tasks:Running task internal.check_lock... 2026-03-21T06:44:02.416 INFO:teuthology.task.internal.check_lock:Checking locks... 2026-03-21T06:44:02.422 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm02.local', 'description': '/archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-21 06:42:20.248126', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:02', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBLcZNeO8u6CEcqUxixY+8HC17tcUZvEGRDk5nT+gHnfF8RoNMc0vTT8W+3sP5EKFyqmgPI7kcILj/IibNmZNa7c='} 2026-03-21T06:44:02.427 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm04.local', 'description': '/archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-21 06:42:20.248925', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:04', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBD2x50LP3vGTd+y7mFC05mi/JRpXP/fcTXHJPS4mJXsdem9ZKd8/nGs2eRti3eHTdUiHgbNHam+5hy3IwloQqaY='} 2026-03-21T06:44:02.431 DEBUG:teuthology.task.internal.check_lock:machine status is {'name': 'vm07.local', 'description': '/archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458', 'up': True, 'machine_type': 'vps', 'is_vm': True, 'vm_host': {'name': 'localhost', 'description': None, 'up': True, 'machine_type': 'libvirt', 'is_vm': False, 'vm_host': None, 'os_type': None, 'os_version': None, 'arch': None, 'locked': True, 'locked_since': None, 'locked_by': None, 'mac_address': None, 'ssh_pub_key': None}, 'os_type': 'ubuntu', 'os_version': '22.04', 'arch': 'x86_64', 'locked': True, 'locked_since': '2026-03-21 06:42:20.248697', 'locked_by': 'kyr', 'mac_address': '52:55:00:00:00:07', 'ssh_pub_key': 'ecdsa-sha2-nistp256 AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBMQCNACJyBC9mJ7lgcIhNI5gd78lRhE4sNx6XNqeQSGilD+wEpJHdUiikPQDbhYkQs5jIcdpYmXGauo70zoDR/k='} 2026-03-21T06:44:02.431 INFO:teuthology.run_tasks:Running task internal.add_remotes... 2026-03-21T06:44:02.431 INFO:teuthology.task.internal:roles: ubuntu@vm02.local - ['host.a', 'mon.a', 'mgr.x', 'osd.0', 'osd.1', 'client.0', 'ceph.iscsi.iscsi.a'] 2026-03-21T06:44:02.431 INFO:teuthology.task.internal:roles: ubuntu@vm04.local - ['mon.b', 'osd.2', 'osd.3', 'osd.4', 'client.1'] 2026-03-21T06:44:02.432 INFO:teuthology.task.internal:roles: ubuntu@vm07.local - ['mon.c', 'osd.5', 'osd.6', 'osd.7', 'client.2', 'ceph.iscsi.iscsi.b'] 2026-03-21T06:44:02.432 INFO:teuthology.run_tasks:Running task console_log... 2026-03-21T06:44:02.437 DEBUG:teuthology.task.console_log:vm02 does not support IPMI; excluding 2026-03-21T06:44:02.442 DEBUG:teuthology.task.console_log:vm04 does not support IPMI; excluding 2026-03-21T06:44:02.446 DEBUG:teuthology.task.console_log:vm07 does not support IPMI; excluding 2026-03-21T06:44:02.446 DEBUG:teuthology.exit:Installing handler: Handler(exiter=, func=.kill_console_loggers at 0x7fe7842089d0>, signals=[15]) 2026-03-21T06:44:02.446 INFO:teuthology.run_tasks:Running task internal.connect... 2026-03-21T06:44:02.447 INFO:teuthology.task.internal:Opening connections... 2026-03-21T06:44:02.447 DEBUG:teuthology.task.internal:connecting to ubuntu@vm02.local 2026-03-21T06:44:02.447 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-21T06:44:02.505 DEBUG:teuthology.task.internal:connecting to ubuntu@vm04.local 2026-03-21T06:44:02.506 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-21T06:44:02.567 DEBUG:teuthology.task.internal:connecting to ubuntu@vm07.local 2026-03-21T06:44:02.567 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm07.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-21T06:44:02.628 INFO:teuthology.run_tasks:Running task internal.push_inventory... 2026-03-21T06:44:02.629 DEBUG:teuthology.orchestra.run.vm02:> uname -m 2026-03-21T06:44:02.632 INFO:teuthology.orchestra.run.vm02.stdout:x86_64 2026-03-21T06:44:02.632 DEBUG:teuthology.orchestra.run.vm02:> cat /etc/os-release 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:NAME="Ubuntu" 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_ID="22.04" 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:VERSION_CODENAME=jammy 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:ID=ubuntu 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:ID_LIKE=debian 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-21T06:44:02.678 INFO:teuthology.orchestra.run.vm02.stdout:UBUNTU_CODENAME=jammy 2026-03-21T06:44:02.678 INFO:teuthology.lock.ops:Updating vm02.local on lock server 2026-03-21T06:44:02.687 DEBUG:teuthology.orchestra.run.vm04:> uname -m 2026-03-21T06:44:02.690 INFO:teuthology.orchestra.run.vm04.stdout:x86_64 2026-03-21T06:44:02.690 DEBUG:teuthology.orchestra.run.vm04:> cat /etc/os-release 2026-03-21T06:44:02.735 INFO:teuthology.orchestra.run.vm04.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-21T06:44:02.736 INFO:teuthology.orchestra.run.vm04.stdout:NAME="Ubuntu" 2026-03-21T06:44:02.736 INFO:teuthology.orchestra.run.vm04.stdout:VERSION_ID="22.04" 2026-03-21T06:44:02.736 INFO:teuthology.orchestra.run.vm04.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-21T06:44:02.736 INFO:teuthology.orchestra.run.vm04.stdout:VERSION_CODENAME=jammy 2026-03-21T06:44:02.736 INFO:teuthology.orchestra.run.vm04.stdout:ID=ubuntu 2026-03-21T06:44:02.736 INFO:teuthology.orchestra.run.vm04.stdout:ID_LIKE=debian 2026-03-21T06:44:02.736 INFO:teuthology.orchestra.run.vm04.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-21T06:44:02.736 INFO:teuthology.orchestra.run.vm04.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-21T06:44:02.736 INFO:teuthology.orchestra.run.vm04.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-21T06:44:02.736 INFO:teuthology.orchestra.run.vm04.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-21T06:44:02.736 INFO:teuthology.orchestra.run.vm04.stdout:UBUNTU_CODENAME=jammy 2026-03-21T06:44:02.736 INFO:teuthology.lock.ops:Updating vm04.local on lock server 2026-03-21T06:44:02.740 DEBUG:teuthology.orchestra.run.vm07:> uname -m 2026-03-21T06:44:02.743 INFO:teuthology.orchestra.run.vm07.stdout:x86_64 2026-03-21T06:44:02.743 DEBUG:teuthology.orchestra.run.vm07:> cat /etc/os-release 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:PRETTY_NAME="Ubuntu 22.04.5 LTS" 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:NAME="Ubuntu" 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:VERSION_ID="22.04" 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:VERSION="22.04.5 LTS (Jammy Jellyfish)" 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:VERSION_CODENAME=jammy 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:ID=ubuntu 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:ID_LIKE=debian 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:HOME_URL="https://www.ubuntu.com/" 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:SUPPORT_URL="https://help.ubuntu.com/" 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" 2026-03-21T06:44:02.793 INFO:teuthology.orchestra.run.vm07.stdout:UBUNTU_CODENAME=jammy 2026-03-21T06:44:02.793 INFO:teuthology.lock.ops:Updating vm07.local on lock server 2026-03-21T06:44:02.797 INFO:teuthology.run_tasks:Running task internal.serialize_remote_roles... 2026-03-21T06:44:02.799 INFO:teuthology.run_tasks:Running task internal.check_conflict... 2026-03-21T06:44:02.800 INFO:teuthology.task.internal:Checking for old test directory... 2026-03-21T06:44:02.800 DEBUG:teuthology.orchestra.run.vm02:> test '!' -e /home/ubuntu/cephtest 2026-03-21T06:44:02.801 DEBUG:teuthology.orchestra.run.vm04:> test '!' -e /home/ubuntu/cephtest 2026-03-21T06:44:02.802 DEBUG:teuthology.orchestra.run.vm07:> test '!' -e /home/ubuntu/cephtest 2026-03-21T06:44:02.837 INFO:teuthology.run_tasks:Running task internal.check_ceph_data... 2026-03-21T06:44:02.838 INFO:teuthology.task.internal:Checking for non-empty /var/lib/ceph... 2026-03-21T06:44:02.838 DEBUG:teuthology.orchestra.run.vm02:> test -z $(ls -A /var/lib/ceph) 2026-03-21T06:44:02.844 DEBUG:teuthology.orchestra.run.vm04:> test -z $(ls -A /var/lib/ceph) 2026-03-21T06:44:02.845 DEBUG:teuthology.orchestra.run.vm07:> test -z $(ls -A /var/lib/ceph) 2026-03-21T06:44:02.846 INFO:teuthology.orchestra.run.vm02.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-21T06:44:02.847 INFO:teuthology.orchestra.run.vm04.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-21T06:44:02.881 INFO:teuthology.orchestra.run.vm07.stderr:ls: cannot access '/var/lib/ceph': No such file or directory 2026-03-21T06:44:02.882 INFO:teuthology.run_tasks:Running task internal.vm_setup... 2026-03-21T06:44:02.889 DEBUG:teuthology.orchestra.run.vm02:> test -e /ceph-qa-ready 2026-03-21T06:44:02.891 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-21T06:44:03.122 DEBUG:teuthology.orchestra.run.vm04:> test -e /ceph-qa-ready 2026-03-21T06:44:03.125 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-21T06:44:03.366 DEBUG:teuthology.orchestra.run.vm07:> test -e /ceph-qa-ready 2026-03-21T06:44:03.368 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-21T06:44:03.590 INFO:teuthology.run_tasks:Running task internal.base... 2026-03-21T06:44:03.591 INFO:teuthology.task.internal:Creating test directory... 2026-03-21T06:44:03.591 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-21T06:44:03.592 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-21T06:44:03.593 DEBUG:teuthology.orchestra.run.vm07:> mkdir -p -m0755 -- /home/ubuntu/cephtest 2026-03-21T06:44:03.596 INFO:teuthology.run_tasks:Running task internal.archive_upload... 2026-03-21T06:44:03.597 INFO:teuthology.run_tasks:Running task internal.archive... 2026-03-21T06:44:03.598 INFO:teuthology.task.internal:Creating archive directory... 2026-03-21T06:44:03.598 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-21T06:44:03.636 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-21T06:44:03.637 DEBUG:teuthology.orchestra.run.vm07:> install -d -m0755 -- /home/ubuntu/cephtest/archive 2026-03-21T06:44:03.643 INFO:teuthology.run_tasks:Running task internal.coredump... 2026-03-21T06:44:03.644 INFO:teuthology.task.internal:Enabling coredump saving... 2026-03-21T06:44:03.644 DEBUG:teuthology.orchestra.run.vm02:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-21T06:44:03.681 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-21T06:44:03.681 DEBUG:teuthology.orchestra.run.vm04:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-21T06:44:03.683 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-21T06:44:03.683 DEBUG:teuthology.orchestra.run.vm07:> test -f /run/.containerenv -o -f /.dockerenv 2026-03-21T06:44:03.685 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-21T06:44:03.685 DEBUG:teuthology.orchestra.run.vm02:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-21T06:44:03.724 DEBUG:teuthology.orchestra.run.vm04:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-21T06:44:03.726 DEBUG:teuthology.orchestra.run.vm07:> install -d -m0755 -- /home/ubuntu/cephtest/archive/coredump && sudo sysctl -w kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core && echo kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core | sudo tee -a /etc/sysctl.conf 2026-03-21T06:44:03.732 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-21T06:44:03.733 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-21T06:44:03.735 INFO:teuthology.orchestra.run.vm07.stdout:kernel.core_pattern = /home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-21T06:44:03.736 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-21T06:44:03.738 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-21T06:44:03.739 INFO:teuthology.orchestra.run.vm07.stdout:kernel.core_pattern=/home/ubuntu/cephtest/archive/coredump/%t.%p.core 2026-03-21T06:44:03.740 INFO:teuthology.run_tasks:Running task internal.sudo... 2026-03-21T06:44:03.741 INFO:teuthology.task.internal:Configuring sudo... 2026-03-21T06:44:03.742 DEBUG:teuthology.orchestra.run.vm02:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-21T06:44:03.780 DEBUG:teuthology.orchestra.run.vm04:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-21T06:44:03.781 DEBUG:teuthology.orchestra.run.vm07:> sudo sed -i.orig.teuthology -e 's/^\([^#]*\) \(requiretty\)/\1 !\2/g' -e 's/^\([^#]*\) !\(visiblepw\)/\1 \2/g' /etc/sudoers 2026-03-21T06:44:03.790 INFO:teuthology.run_tasks:Running task internal.syslog... 2026-03-21T06:44:03.793 INFO:teuthology.task.internal.syslog:Starting syslog monitoring... 2026-03-21T06:44:03.793 DEBUG:teuthology.orchestra.run.vm02:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-21T06:44:03.828 DEBUG:teuthology.orchestra.run.vm04:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-21T06:44:03.830 DEBUG:teuthology.orchestra.run.vm07:> mkdir -p -m0755 -- /home/ubuntu/cephtest/archive/syslog 2026-03-21T06:44:03.833 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-21T06:44:03.878 DEBUG:teuthology.orchestra.run.vm02:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-21T06:44:03.922 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:44:03.922 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-21T06:44:03.971 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-21T06:44:03.975 DEBUG:teuthology.orchestra.run.vm04:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-21T06:44:04.020 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:44:04.020 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-21T06:44:04.072 DEBUG:teuthology.orchestra.run.vm07:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-21T06:44:04.076 DEBUG:teuthology.orchestra.run.vm07:> install -m 666 /dev/null /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-21T06:44:04.121 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:44:04.121 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/rsyslog.d/80-cephtest.conf 2026-03-21T06:44:04.170 DEBUG:teuthology.orchestra.run.vm02:> sudo service rsyslog restart 2026-03-21T06:44:04.170 DEBUG:teuthology.orchestra.run.vm04:> sudo service rsyslog restart 2026-03-21T06:44:04.171 DEBUG:teuthology.orchestra.run.vm07:> sudo service rsyslog restart 2026-03-21T06:44:04.225 INFO:teuthology.run_tasks:Running task internal.timer... 2026-03-21T06:44:04.227 INFO:teuthology.task.internal:Starting timer... 2026-03-21T06:44:04.227 INFO:teuthology.run_tasks:Running task pcp... 2026-03-21T06:44:04.229 INFO:teuthology.run_tasks:Running task selinux... 2026-03-21T06:44:04.231 INFO:teuthology.task.selinux:Excluding vm02: VMs are not yet supported 2026-03-21T06:44:04.232 INFO:teuthology.task.selinux:Excluding vm04: VMs are not yet supported 2026-03-21T06:44:04.232 INFO:teuthology.task.selinux:Excluding vm07: VMs are not yet supported 2026-03-21T06:44:04.232 DEBUG:teuthology.task.selinux:Getting current SELinux state 2026-03-21T06:44:04.232 DEBUG:teuthology.task.selinux:Existing SELinux modes: {} 2026-03-21T06:44:04.232 INFO:teuthology.task.selinux:Putting SELinux into permissive mode 2026-03-21T06:44:04.232 INFO:teuthology.run_tasks:Running task ansible.cephlab... 2026-03-21T06:44:04.233 DEBUG:teuthology.task:Applying overrides for task ansible.cephlab: {'branch': 'main', 'repo': 'https://github.com/kshtsk/ceph-cm-ansible.git', 'skip_tags': 'nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs', 'vars': {'logical_volumes': {'lv_1': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_2': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_3': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}, 'lv_4': {'scratch_dev': True, 'size': '25%VG', 'vg': 'vg_nvme'}}, 'timezone': 'UTC', 'volume_groups': {'vg_nvme': {'pvs': '/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde'}}}} 2026-03-21T06:44:04.233 DEBUG:teuthology.repo_utils:Setting repo remote to https://github.com/kshtsk/ceph-cm-ansible.git 2026-03-21T06:44:04.238 INFO:teuthology.repo_utils:Fetching github.com_kshtsk_ceph-cm-ansible_main from origin 2026-03-21T06:44:04.821 DEBUG:teuthology.repo_utils:Resetting repo at /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main to origin/main 2026-03-21T06:44:04.826 INFO:teuthology.task.ansible:Playbook: [{'import_playbook': 'ansible_managed.yml'}, {'import_playbook': 'teuthology.yml'}, {'hosts': 'testnodes', 'tasks': [{'set_fact': {'ran_from_cephlab_playbook': True}}]}, {'import_playbook': 'testnodes.yml'}, {'import_playbook': 'container-host.yml'}, {'import_playbook': 'cobbler.yml'}, {'import_playbook': 'paddles.yml'}, {'import_playbook': 'pulpito.yml'}, {'hosts': 'testnodes', 'become': True, 'tasks': [{'name': 'Touch /ceph-qa-ready', 'file': {'path': '/ceph-qa-ready', 'state': 'touch'}, 'when': 'ran_from_cephlab_playbook|bool'}]}] 2026-03-21T06:44:04.826 DEBUG:teuthology.task.ansible:Running ansible-playbook -v --extra-vars '{"ansible_ssh_user": "ubuntu", "logical_volumes": {"lv_1": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_2": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_3": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}, "lv_4": {"scratch_dev": true, "size": "25%VG", "vg": "vg_nvme"}}, "timezone": "UTC", "volume_groups": {"vg_nvme": {"pvs": "/dev/vdb,/dev/vdc,/dev/vdd,/dev/vde"}}}' -i /tmp/teuth_ansible_inventory0hi2cqw2 --limit vm02.local,vm04.local,vm07.local /home/teuthos/src/github.com_kshtsk_ceph-cm-ansible_main/cephlab.yml --skip-tags nagios,monitoring-scripts,hostname,pubkeys,zap,sudoers,kerberos,ntp-client,resolvconf,cpan,nfs 2026-03-21T06:46:20.857 DEBUG:teuthology.task.ansible:Reconnecting to [Remote(name='ubuntu@vm02.local'), Remote(name='ubuntu@vm04.local'), Remote(name='ubuntu@vm07.local')] 2026-03-21T06:46:20.857 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm02.local' 2026-03-21T06:46:20.858 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm02.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-21T06:46:20.917 DEBUG:teuthology.orchestra.run.vm02:> true 2026-03-21T06:46:21.116 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm02.local' 2026-03-21T06:46:21.116 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm04.local' 2026-03-21T06:46:21.116 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm04.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-21T06:46:21.175 DEBUG:teuthology.orchestra.run.vm04:> true 2026-03-21T06:46:21.384 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm04.local' 2026-03-21T06:46:21.384 INFO:teuthology.orchestra.remote:Trying to reconnect to host 'ubuntu@vm07.local' 2026-03-21T06:46:21.384 DEBUG:teuthology.orchestra.connection:{'hostname': 'vm07.local', 'username': 'ubuntu', 'timeout': 60} 2026-03-21T06:46:21.443 DEBUG:teuthology.orchestra.run.vm07:> true 2026-03-21T06:46:21.644 INFO:teuthology.orchestra.remote:Successfully reconnected to host 'ubuntu@vm07.local' 2026-03-21T06:46:21.644 INFO:teuthology.run_tasks:Running task clock... 2026-03-21T06:46:21.646 INFO:teuthology.task.clock:Syncing clocks and checking initial clock skew... 2026-03-21T06:46:21.647 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-21T06:46:21.647 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-21T06:46:21.648 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-21T06:46:21.648 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-21T06:46:21.649 INFO:teuthology.orchestra.run:Running command with timeout 360 2026-03-21T06:46:21.649 DEBUG:teuthology.orchestra.run.vm07:> sudo systemctl stop ntp.service || sudo systemctl stop ntpd.service || sudo systemctl stop chronyd.service ; sudo ntpd -gq || sudo chronyc makestep ; sudo systemctl start ntp.service || sudo systemctl start ntpd.service || sudo systemctl start chronyd.service ; PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-21T06:46:21.662 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-21T06:46:21.662 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: Command line: ntpd -gq 2026-03-21T06:46:21.662 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: ---------------------------------------------------- 2026-03-21T06:46:21.662 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: ntp-4 is maintained by Network Time Foundation, 2026-03-21T06:46:21.662 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-21T06:46:21.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: corporation. Support and training for ntp-4 are 2026-03-21T06:46:21.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: available at https://www.nwtime.org/support 2026-03-21T06:46:21.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: ---------------------------------------------------- 2026-03-21T06:46:21.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: proto: precision = 0.029 usec (-25) 2026-03-21T06:46:21.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: basedate set to 2022-02-04 2026-03-21T06:46:21.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: gps base set to 2022-02-06 (week 2196) 2026-03-21T06:46:21.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-21T06:46:21.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-21T06:46:21.663 INFO:teuthology.orchestra.run.vm02.stderr:21 Mar 06:46:21 ntpd[16248]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 84 days ago 2026-03-21T06:46:21.664 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: Listen and drop on 0 v6wildcard [::]:123 2026-03-21T06:46:21.664 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-21T06:46:21.664 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: Listen normally on 2 lo 127.0.0.1:123 2026-03-21T06:46:21.664 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: Listen normally on 3 ens3 192.168.123.102:123 2026-03-21T06:46:21.664 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: Listen normally on 4 lo [::1]:123 2026-03-21T06:46:21.664 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:2%2]:123 2026-03-21T06:46:21.664 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:21 ntpd[16248]: Listening on routing socket on fd #22 for interface updates 2026-03-21T06:46:21.666 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-21T06:46:21.666 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: Command line: ntpd -gq 2026-03-21T06:46:21.666 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: ---------------------------------------------------- 2026-03-21T06:46:21.666 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: ntp-4 is maintained by Network Time Foundation, 2026-03-21T06:46:21.666 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-21T06:46:21.666 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: corporation. Support and training for ntp-4 are 2026-03-21T06:46:21.666 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: available at https://www.nwtime.org/support 2026-03-21T06:46:21.666 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: ---------------------------------------------------- 2026-03-21T06:46:21.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: proto: precision = 0.030 usec (-25) 2026-03-21T06:46:21.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: basedate set to 2022-02-04 2026-03-21T06:46:21.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: gps base set to 2022-02-06 (week 2196) 2026-03-21T06:46:21.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-21T06:46:21.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-21T06:46:21.667 INFO:teuthology.orchestra.run.vm04.stderr:21 Mar 06:46:21 ntpd[16229]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 84 days ago 2026-03-21T06:46:21.668 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: Listen and drop on 0 v6wildcard [::]:123 2026-03-21T06:46:21.668 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-21T06:46:21.668 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: Listen normally on 2 lo 127.0.0.1:123 2026-03-21T06:46:21.668 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: Listen normally on 3 ens3 192.168.123.104:123 2026-03-21T06:46:21.668 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: Listen normally on 4 lo [::1]:123 2026-03-21T06:46:21.668 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:4%2]:123 2026-03-21T06:46:21.668 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:21 ntpd[16229]: Listening on routing socket on fd #22 for interface updates 2026-03-21T06:46:21.702 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: ntpd 4.2.8p15@1.3728-o Wed Feb 16 17:13:02 UTC 2022 (1): Starting 2026-03-21T06:46:21.702 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: Command line: ntpd -gq 2026-03-21T06:46:21.702 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: ---------------------------------------------------- 2026-03-21T06:46:21.702 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: ntp-4 is maintained by Network Time Foundation, 2026-03-21T06:46:21.702 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: Inc. (NTF), a non-profit 501(c)(3) public-benefit 2026-03-21T06:46:21.702 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: corporation. Support and training for ntp-4 are 2026-03-21T06:46:21.702 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: available at https://www.nwtime.org/support 2026-03-21T06:46:21.702 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: ---------------------------------------------------- 2026-03-21T06:46:21.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: proto: precision = 0.030 usec (-25) 2026-03-21T06:46:21.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: basedate set to 2022-02-04 2026-03-21T06:46:21.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: gps base set to 2022-02-06 (week 2196) 2026-03-21T06:46:21.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): good hash signature 2026-03-21T06:46:21.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): loaded, expire=2025-12-28T00:00:00Z last=2017-01-01T00:00:00Z ofs=37 2026-03-21T06:46:21.703 INFO:teuthology.orchestra.run.vm07.stderr:21 Mar 06:46:21 ntpd[16233]: leapsecond file ('/usr/share/zoneinfo/leap-seconds.list'): expired 84 days ago 2026-03-21T06:46:21.704 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: Listen and drop on 0 v6wildcard [::]:123 2026-03-21T06:46:21.704 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: Listen and drop on 1 v4wildcard 0.0.0.0:123 2026-03-21T06:46:21.704 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: Listen normally on 2 lo 127.0.0.1:123 2026-03-21T06:46:21.704 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: Listen normally on 3 ens3 192.168.123.107:123 2026-03-21T06:46:21.704 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: Listen normally on 4 lo [::1]:123 2026-03-21T06:46:21.704 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: Listen normally on 5 ens3 [fe80::5055:ff:fe00:7%2]:123 2026-03-21T06:46:21.705 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:21 ntpd[16233]: Listening on routing socket on fd #22 for interface updates 2026-03-21T06:46:22.664 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:22 ntpd[16248]: Soliciting pool server 178.63.9.212 2026-03-21T06:46:22.668 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:22 ntpd[16229]: Soliciting pool server 178.63.9.212 2026-03-21T06:46:22.704 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:22 ntpd[16233]: Soliciting pool server 173.249.58.145 2026-03-21T06:46:23.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:23 ntpd[16248]: Soliciting pool server 51.75.67.47 2026-03-21T06:46:23.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:23 ntpd[16248]: Soliciting pool server 31.209.85.243 2026-03-21T06:46:23.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:23 ntpd[16229]: Soliciting pool server 31.209.85.243 2026-03-21T06:46:23.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:23 ntpd[16233]: Soliciting pool server 178.63.9.212 2026-03-21T06:46:23.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:23 ntpd[16233]: Soliciting pool server 157.230.22.48 2026-03-21T06:46:24.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:24 ntpd[16248]: Soliciting pool server 159.69.64.189 2026-03-21T06:46:24.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:24 ntpd[16248]: Soliciting pool server 78.47.56.71 2026-03-21T06:46:24.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:24 ntpd[16248]: Soliciting pool server 162.55.190.98 2026-03-21T06:46:24.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:24 ntpd[16229]: Soliciting pool server 159.69.64.189 2026-03-21T06:46:24.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:24 ntpd[16229]: Soliciting pool server 162.55.190.98 2026-03-21T06:46:24.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:24 ntpd[16233]: Soliciting pool server 31.209.85.243 2026-03-21T06:46:24.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:24 ntpd[16233]: Soliciting pool server 185.232.69.65 2026-03-21T06:46:25.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:25 ntpd[16248]: Soliciting pool server 5.45.97.204 2026-03-21T06:46:25.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:25 ntpd[16248]: Soliciting pool server 144.76.59.37 2026-03-21T06:46:25.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:25 ntpd[16248]: Soliciting pool server 173.249.58.145 2026-03-21T06:46:25.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:25 ntpd[16248]: Soliciting pool server 162.159.200.123 2026-03-21T06:46:25.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:25 ntpd[16229]: Soliciting pool server 5.45.97.204 2026-03-21T06:46:25.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:25 ntpd[16229]: Soliciting pool server 144.76.59.37 2026-03-21T06:46:25.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:25 ntpd[16229]: Soliciting pool server 162.159.200.123 2026-03-21T06:46:25.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:25 ntpd[16233]: Soliciting pool server 162.55.190.98 2026-03-21T06:46:25.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:25 ntpd[16233]: Soliciting pool server 159.69.64.189 2026-03-21T06:46:25.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:25 ntpd[16233]: Soliciting pool server 176.9.157.155 2026-03-21T06:46:26.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:26 ntpd[16248]: Soliciting pool server 139.162.156.95 2026-03-21T06:46:26.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:26 ntpd[16248]: Soliciting pool server 85.10.240.253 2026-03-21T06:46:26.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:26 ntpd[16248]: Soliciting pool server 157.230.22.48 2026-03-21T06:46:26.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:26 ntpd[16248]: Soliciting pool server 91.189.91.157 2026-03-21T06:46:26.666 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:26 ntpd[16229]: Soliciting pool server 139.162.156.95 2026-03-21T06:46:26.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:26 ntpd[16229]: Soliciting pool server 85.10.240.253 2026-03-21T06:46:26.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:26 ntpd[16229]: Soliciting pool server 157.230.22.48 2026-03-21T06:46:26.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:26 ntpd[16229]: Soliciting pool server 91.189.91.157 2026-03-21T06:46:26.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:26 ntpd[16233]: Soliciting pool server 162.159.200.123 2026-03-21T06:46:26.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:26 ntpd[16233]: Soliciting pool server 5.45.97.204 2026-03-21T06:46:26.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:26 ntpd[16233]: Soliciting pool server 144.76.59.37 2026-03-21T06:46:26.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:26 ntpd[16233]: Soliciting pool server 185.125.190.58 2026-03-21T06:46:27.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:27 ntpd[16248]: Soliciting pool server 185.125.190.57 2026-03-21T06:46:27.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:27 ntpd[16248]: Soliciting pool server 157.90.247.99 2026-03-21T06:46:27.663 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:27 ntpd[16248]: Soliciting pool server 185.232.69.65 2026-03-21T06:46:27.664 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:27 ntpd[16248]: Soliciting pool server 5.75.181.179 2026-03-21T06:46:27.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:27 ntpd[16229]: Soliciting pool server 185.125.190.57 2026-03-21T06:46:27.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:27 ntpd[16229]: Soliciting pool server 157.90.247.99 2026-03-21T06:46:27.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:27 ntpd[16229]: Soliciting pool server 185.232.69.65 2026-03-21T06:46:27.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:27 ntpd[16229]: Soliciting pool server 5.75.181.179 2026-03-21T06:46:27.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:27 ntpd[16233]: Soliciting pool server 91.189.91.157 2026-03-21T06:46:27.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:27 ntpd[16233]: Soliciting pool server 139.162.156.95 2026-03-21T06:46:27.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:27 ntpd[16233]: Soliciting pool server 85.10.240.253 2026-03-21T06:46:27.704 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:27 ntpd[16233]: Soliciting pool server 91.198.32.34 2026-03-21T06:46:28.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:28 ntpd[16229]: Soliciting pool server 185.125.190.56 2026-03-21T06:46:28.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:28 ntpd[16229]: Soliciting pool server 176.9.157.155 2026-03-21T06:46:28.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:28 ntpd[16229]: Soliciting pool server 2a01:4f8:150:60c3::2 2026-03-21T06:46:28.667 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:28 ntpd[16229]: Soliciting pool server 64.188.67.160 2026-03-21T06:46:28.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:28 ntpd[16233]: Soliciting pool server 185.125.190.57 2026-03-21T06:46:28.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:28 ntpd[16233]: Soliciting pool server 157.90.247.99 2026-03-21T06:46:28.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:28 ntpd[16233]: Soliciting pool server 2003:a:843:7c00::3 2026-03-21T06:46:28.703 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:28 ntpd[16233]: Soliciting pool server 5.75.181.179 2026-03-21T06:46:29.739 INFO:teuthology.orchestra.run.vm07.stdout:21 Mar 06:46:29 ntpd[16233]: ntpd: time slew +0.004509 s 2026-03-21T06:46:29.739 INFO:teuthology.orchestra.run.vm07.stdout:ntpd: time slew +0.004509s 2026-03-21T06:46:29.761 INFO:teuthology.orchestra.run.vm07.stdout: remote refid st t when poll reach delay offset jitter 2026-03-21T06:46:29.761 INFO:teuthology.orchestra.run.vm07.stdout:============================================================================== 2026-03-21T06:46:29.761 INFO:teuthology.orchestra.run.vm07.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:29.762 INFO:teuthology.orchestra.run.vm07.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:29.762 INFO:teuthology.orchestra.run.vm07.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:29.762 INFO:teuthology.orchestra.run.vm07.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:29.762 INFO:teuthology.orchestra.run.vm07.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:30.685 INFO:teuthology.orchestra.run.vm02.stdout:21 Mar 06:46:30 ntpd[16248]: ntpd: time slew +0.001573 s 2026-03-21T06:46:30.685 INFO:teuthology.orchestra.run.vm02.stdout:ntpd: time slew +0.001573s 2026-03-21T06:46:30.698 INFO:teuthology.orchestra.run.vm04.stdout:21 Mar 06:46:30 ntpd[16229]: ntpd: time slew +0.004540 s 2026-03-21T06:46:30.698 INFO:teuthology.orchestra.run.vm04.stdout:ntpd: time slew +0.004540s 2026-03-21T06:46:30.708 INFO:teuthology.orchestra.run.vm02.stdout: remote refid st t when poll reach delay offset jitter 2026-03-21T06:46:30.708 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================== 2026-03-21T06:46:30.708 INFO:teuthology.orchestra.run.vm02.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:30.708 INFO:teuthology.orchestra.run.vm02.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:30.708 INFO:teuthology.orchestra.run.vm02.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:30.708 INFO:teuthology.orchestra.run.vm02.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:30.708 INFO:teuthology.orchestra.run.vm02.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:30.725 INFO:teuthology.orchestra.run.vm04.stdout: remote refid st t when poll reach delay offset jitter 2026-03-21T06:46:30.725 INFO:teuthology.orchestra.run.vm04.stdout:============================================================================== 2026-03-21T06:46:30.725 INFO:teuthology.orchestra.run.vm04.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:30.725 INFO:teuthology.orchestra.run.vm04.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:30.725 INFO:teuthology.orchestra.run.vm04.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:30.725 INFO:teuthology.orchestra.run.vm04.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:30.725 INFO:teuthology.orchestra.run.vm04.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T06:46:30.726 INFO:teuthology.run_tasks:Running task cephadm... 2026-03-21T06:46:30.774 INFO:tasks.cephadm:Config: {'conf': {'global': {'mon warn on pool no app': False}, 'mgr': {'debug mgr': 20, 'debug ms': 1}, 'mon': {'debug mon': 20, 'debug ms': 1, 'debug paxos': 20}, 'osd': {'debug ms': 1, 'debug osd': 20, 'osd mclock iops capacity threshold hdd': 49000}}, 'flavor': 'default', 'log-ignorelist': ['\\(MDS_ALL_DOWN\\)', '\\(MDS_UP_LESS_THAN_MAX\\)', 'MON_DOWN'], 'sha1': '70f8415b300f041766fa27faf7d5472699e32388', 'cephadm_binary_url': 'https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm'} 2026-03-21T06:46:30.774 INFO:tasks.cephadm:Cluster image is quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 2026-03-21T06:46:30.774 INFO:tasks.cephadm:Cluster fsid is b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:46:30.774 INFO:tasks.cephadm:Choosing monitor IPs and ports... 2026-03-21T06:46:30.774 INFO:tasks.cephadm:Monitor IPs: {'mon.a': '192.168.123.102', 'mon.b': '192.168.123.104', 'mon.c': '192.168.123.107'} 2026-03-21T06:46:30.774 INFO:tasks.cephadm:First mon is mon.a on vm02 2026-03-21T06:46:30.774 INFO:tasks.cephadm:First mgr is x 2026-03-21T06:46:30.774 INFO:tasks.cephadm:Normalizing hostnames... 2026-03-21T06:46:30.774 DEBUG:teuthology.orchestra.run.vm02:> sudo hostname $(hostname -s) 2026-03-21T06:46:30.783 DEBUG:teuthology.orchestra.run.vm04:> sudo hostname $(hostname -s) 2026-03-21T06:46:30.791 DEBUG:teuthology.orchestra.run.vm07:> sudo hostname $(hostname -s) 2026-03-21T06:46:30.800 INFO:tasks.cephadm:Downloading cephadm from url: https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm 2026-03-21T06:46:30.801 DEBUG:teuthology.orchestra.run.vm02:> curl --silent -L https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-21T06:46:32.737 INFO:teuthology.orchestra.run.vm02.stdout:-rw-rw-r-- 1 ubuntu ubuntu 1036391 Mar 21 06:46 /home/ubuntu/cephtest/cephadm 2026-03-21T06:46:32.738 DEBUG:teuthology.orchestra.run.vm04:> curl --silent -L https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-21T06:46:37.358 INFO:teuthology.orchestra.run.vm04.stdout:-rw-rw-r-- 1 ubuntu ubuntu 1036391 Mar 21 06:46 /home/ubuntu/cephtest/cephadm 2026-03-21T06:46:37.359 DEBUG:teuthology.orchestra.run.vm07:> curl --silent -L https://download.ceph.com/rpm-20.2.0/el9/noarch/cephadm > /home/ubuntu/cephtest/cephadm && ls -l /home/ubuntu/cephtest/cephadm 2026-03-21T06:46:40.637 INFO:teuthology.orchestra.run.vm07.stdout:-rw-rw-r-- 1 ubuntu ubuntu 1036391 Mar 21 06:46 /home/ubuntu/cephtest/cephadm 2026-03-21T06:46:40.637 DEBUG:teuthology.orchestra.run.vm02:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-21T06:46:40.641 DEBUG:teuthology.orchestra.run.vm04:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-21T06:46:40.647 DEBUG:teuthology.orchestra.run.vm07:> test -s /home/ubuntu/cephtest/cephadm && test $(stat -c%s /home/ubuntu/cephtest/cephadm) -gt 1000 && chmod +x /home/ubuntu/cephtest/cephadm 2026-03-21T06:46:40.655 INFO:tasks.cephadm:Pulling image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 on all hosts... 2026-03-21T06:46:40.655 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 pull 2026-03-21T06:46:40.683 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 pull 2026-03-21T06:46:40.689 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 pull 2026-03-21T06:46:40.963 INFO:teuthology.orchestra.run.vm02.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388... 2026-03-21T06:46:40.964 INFO:teuthology.orchestra.run.vm04.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388... 2026-03-21T06:46:40.987 INFO:teuthology.orchestra.run.vm07.stderr:Pulling container image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388... 2026-03-21T06:47:35.201 INFO:teuthology.orchestra.run.vm02.stdout:{ 2026-03-21T06:47:35.201 INFO:teuthology.orchestra.run.vm02.stdout: "ceph_version": "ceph version 20.2.0-712-g70f8415b (70f8415b300f041766fa27faf7d5472699e32388) tentacle (stable)", 2026-03-21T06:47:35.201 INFO:teuthology.orchestra.run.vm02.stdout: "image_id": "26d640ab1cebc48c7b5e1402f96cc470913dba3fe05c2cf5175ea3cde51a580c", 2026-03-21T06:47:35.201 INFO:teuthology.orchestra.run.vm02.stdout: "repo_digests": [ 2026-03-21T06:47:35.201 INFO:teuthology.orchestra.run.vm02.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:5fe4cfc092321d85e013fc971c0a15337be7ad4a3c8cc47ec11af07596687bd6" 2026-03-21T06:47:35.201 INFO:teuthology.orchestra.run.vm02.stdout: ] 2026-03-21T06:47:35.201 INFO:teuthology.orchestra.run.vm02.stdout:} 2026-03-21T06:47:35.226 INFO:teuthology.orchestra.run.vm04.stdout:{ 2026-03-21T06:47:35.226 INFO:teuthology.orchestra.run.vm04.stdout: "ceph_version": "ceph version 20.2.0-712-g70f8415b (70f8415b300f041766fa27faf7d5472699e32388) tentacle (stable)", 2026-03-21T06:47:35.227 INFO:teuthology.orchestra.run.vm04.stdout: "image_id": "26d640ab1cebc48c7b5e1402f96cc470913dba3fe05c2cf5175ea3cde51a580c", 2026-03-21T06:47:35.227 INFO:teuthology.orchestra.run.vm04.stdout: "repo_digests": [ 2026-03-21T06:47:35.227 INFO:teuthology.orchestra.run.vm04.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:5fe4cfc092321d85e013fc971c0a15337be7ad4a3c8cc47ec11af07596687bd6" 2026-03-21T06:47:35.227 INFO:teuthology.orchestra.run.vm04.stdout: ] 2026-03-21T06:47:35.227 INFO:teuthology.orchestra.run.vm04.stdout:} 2026-03-21T06:47:44.309 INFO:teuthology.orchestra.run.vm07.stdout:{ 2026-03-21T06:47:44.309 INFO:teuthology.orchestra.run.vm07.stdout: "ceph_version": "ceph version 20.2.0-712-g70f8415b (70f8415b300f041766fa27faf7d5472699e32388) tentacle (stable)", 2026-03-21T06:47:44.309 INFO:teuthology.orchestra.run.vm07.stdout: "image_id": "26d640ab1cebc48c7b5e1402f96cc470913dba3fe05c2cf5175ea3cde51a580c", 2026-03-21T06:47:44.309 INFO:teuthology.orchestra.run.vm07.stdout: "repo_digests": [ 2026-03-21T06:47:44.309 INFO:teuthology.orchestra.run.vm07.stdout: "quay.ceph.io/ceph-ci/ceph@sha256:5fe4cfc092321d85e013fc971c0a15337be7ad4a3c8cc47ec11af07596687bd6" 2026-03-21T06:47:44.309 INFO:teuthology.orchestra.run.vm07.stdout: ] 2026-03-21T06:47:44.309 INFO:teuthology.orchestra.run.vm07.stdout:} 2026-03-21T06:47:44.325 DEBUG:teuthology.orchestra.run.vm02:> sudo mkdir -p /etc/ceph 2026-03-21T06:47:44.333 DEBUG:teuthology.orchestra.run.vm04:> sudo mkdir -p /etc/ceph 2026-03-21T06:47:44.341 DEBUG:teuthology.orchestra.run.vm07:> sudo mkdir -p /etc/ceph 2026-03-21T06:47:44.348 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 777 /etc/ceph 2026-03-21T06:47:44.382 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod 777 /etc/ceph 2026-03-21T06:47:44.391 DEBUG:teuthology.orchestra.run.vm07:> sudo chmod 777 /etc/ceph 2026-03-21T06:47:44.399 INFO:tasks.cephadm:Writing seed config... 2026-03-21T06:47:44.399 INFO:tasks.cephadm: override: [global] mon warn on pool no app = False 2026-03-21T06:47:44.399 INFO:tasks.cephadm: override: [mgr] debug mgr = 20 2026-03-21T06:47:44.399 INFO:tasks.cephadm: override: [mgr] debug ms = 1 2026-03-21T06:47:44.400 INFO:tasks.cephadm: override: [mon] debug mon = 20 2026-03-21T06:47:44.400 INFO:tasks.cephadm: override: [mon] debug ms = 1 2026-03-21T06:47:44.400 INFO:tasks.cephadm: override: [mon] debug paxos = 20 2026-03-21T06:47:44.400 INFO:tasks.cephadm: override: [osd] debug ms = 1 2026-03-21T06:47:44.400 INFO:tasks.cephadm: override: [osd] debug osd = 20 2026-03-21T06:47:44.400 INFO:tasks.cephadm: override: [osd] osd mclock iops capacity threshold hdd = 49000 2026-03-21T06:47:44.400 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:47:44.400 DEBUG:teuthology.orchestra.run.vm02:> dd of=/home/ubuntu/cephtest/seed.ceph.conf 2026-03-21T06:47:44.426 DEBUG:tasks.cephadm:Final config: [global] # make logging friendly to teuthology log_to_file = true log_to_stderr = false log to journald = false mon cluster log to file = true mon cluster log file level = debug mon clock drift allowed = 1.000 # replicate across OSDs, not hosts osd crush chooseleaf type = 0 #osd pool default size = 2 osd pool default erasure code profile = plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd # enable some debugging auth debug = true ms die on old message = true ms die on bug = true debug asserts on shutdown = true # adjust warnings mon max pg per osd = 10000# >= luminous mon pg warn max object skew = 0 mon osd allow primary affinity = true mon osd allow pg remap = true mon warn on legacy crush tunables = false mon warn on crush straw calc version zero = false mon warn on no sortbitwise = false mon warn on osd down out interval zero = false mon warn on too few osds = false mon_warn_on_pool_pg_num_not_power_of_two = false # disable pg_autoscaler by default for new pools osd_pool_default_pg_autoscale_mode = off # tests delete pools mon allow pool delete = true fsid = b16ecafc-24f1-11f1-8ede-8330751617ee mon warn on pool no app = False [osd] osd scrub load threshold = 5.0 osd scrub max interval = 600 osd mclock profile = high_recovery_ops osd mclock skip benchmark = true osd recover clone overlap = true osd recovery max chunk = 1048576 osd deep scrub update digest min age = 30 osd map max advance = 10 osd memory target autotune = true # debugging osd debug shutdown = true osd debug op order = true osd debug verify stray on activate = true osd debug pg log writeout = true osd debug verify cached snaps = true osd debug verify missing on start = true osd debug misdirected ops = true osd op queue = debug_random osd op queue cut off = debug_random osd shutdown pgref assert = true bdev debug aio = true osd sloppy crc = true debug ms = 1 debug osd = 20 osd mclock iops capacity threshold hdd = 49000 [mgr] mon reweight min pgs per osd = 4 mon reweight min bytes per osd = 10 mgr/telemetry/nag = false debug mgr = 20 debug ms = 1 [mon] mon data avail warn = 5 mon mgr mkfs grace = 240 mon reweight min pgs per osd = 4 mon osd reporter subtree level = osd mon osd prime pg temp = true mon reweight min bytes per osd = 10 # rotate auth tickets quickly to exercise renewal paths auth mon ticket ttl = 660# 11m auth service ticket ttl = 240# 4m # don't complain about global id reclaim mon_warn_on_insecure_global_id_reclaim = false mon_warn_on_insecure_global_id_reclaim_allowed = false debug mon = 20 debug ms = 1 debug paxos = 20 [client.rgw] rgw cache enabled = true rgw enable ops log = true rgw enable usage log = true 2026-03-21T06:47:44.426 DEBUG:teuthology.orchestra.run.vm02:mon.a> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.a.service 2026-03-21T06:47:44.468 DEBUG:teuthology.orchestra.run.vm02:mgr.x> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mgr.x.service 2026-03-21T06:47:44.512 INFO:tasks.cephadm:Bootstrapping... 2026-03-21T06:47:44.512 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 -v bootstrap --fsid b16ecafc-24f1-11f1-8ede-8330751617ee --config /home/ubuntu/cephtest/seed.ceph.conf --output-config /etc/ceph/ceph.conf --output-keyring /etc/ceph/ceph.client.admin.keyring --output-pub-ssh-key /home/ubuntu/cephtest/ceph.pub --mon-id a --mgr-id x --orphan-initial-daemons --skip-monitoring-stack --mon-ip 192.168.123.102 --skip-admin-label && sudo chmod +r /etc/ceph/ceph.client.admin.keyring 2026-03-21T06:47:44.785 INFO:teuthology.orchestra.run.vm02.stdout:-------------------------------------------------------------------------------- 2026-03-21T06:47:44.785 INFO:teuthology.orchestra.run.vm02.stdout:cephadm ['--image', 'quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388', '-v', 'bootstrap', '--fsid', 'b16ecafc-24f1-11f1-8ede-8330751617ee', '--config', '/home/ubuntu/cephtest/seed.ceph.conf', '--output-config', '/etc/ceph/ceph.conf', '--output-keyring', '/etc/ceph/ceph.client.admin.keyring', '--output-pub-ssh-key', '/home/ubuntu/cephtest/ceph.pub', '--mon-id', 'a', '--mgr-id', 'x', '--orphan-initial-daemons', '--skip-monitoring-stack', '--mon-ip', '192.168.123.102', '--skip-admin-label'] 2026-03-21T06:47:44.785 INFO:teuthology.orchestra.run.vm02.stderr:Specifying an fsid for your cluster offers no advantages and may increase the likelihood of fsid conflicts. 2026-03-21T06:47:44.785 INFO:teuthology.orchestra.run.vm02.stdout:Verifying podman|docker is present... 2026-03-21T06:47:44.785 INFO:teuthology.orchestra.run.vm02.stdout:Verifying lvm2 is present... 2026-03-21T06:47:44.785 INFO:teuthology.orchestra.run.vm02.stdout:Verifying time synchronization is in place... 2026-03-21T06:47:44.788 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-21T06:47:44.788 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-21T06:47:44.790 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-21T06:47:44.790 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-21T06:47:44.793 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-21T06:47:44.793 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-21T06:47:44.795 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-21T06:47:44.795 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-21T06:47:44.797 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-21T06:47:44.797 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout masked 2026-03-21T06:47:44.799 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-21T06:47:44.799 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-21T06:47:44.802 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-21T06:47:44.802 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-21T06:47:44.804 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-21T06:47:44.804 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-21T06:47:44.807 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout enabled 2026-03-21T06:47:44.809 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout active 2026-03-21T06:47:44.809 INFO:teuthology.orchestra.run.vm02.stdout:Unit ntp.service is enabled and running 2026-03-21T06:47:44.809 INFO:teuthology.orchestra.run.vm02.stdout:Repeating the final host check... 2026-03-21T06:47:44.809 INFO:teuthology.orchestra.run.vm02.stdout:docker (/usr/bin/docker) is present 2026-03-21T06:47:44.809 INFO:teuthology.orchestra.run.vm02.stdout:systemctl is present 2026-03-21T06:47:44.809 INFO:teuthology.orchestra.run.vm02.stdout:lvcreate is present 2026-03-21T06:47:44.812 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chrony.service 2026-03-21T06:47:44.812 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chrony.service: No such file or directory 2026-03-21T06:47:44.814 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chrony.service 2026-03-21T06:47:44.814 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-21T06:47:44.816 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled chronyd.service 2026-03-21T06:47:44.816 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for chronyd.service: No such file or directory 2026-03-21T06:47:44.818 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active chronyd.service 2026-03-21T06:47:44.818 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-21T06:47:44.821 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled systemd-timesyncd.service 2026-03-21T06:47:44.821 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout masked 2026-03-21T06:47:44.823 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active systemd-timesyncd.service 2026-03-21T06:47:44.823 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-21T06:47:44.825 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl is-enabled ntpd.service 2026-03-21T06:47:44.825 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to get unit file state for ntpd.service: No such file or directory 2026-03-21T06:47:44.828 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active ntpd.service 2026-03-21T06:47:44.828 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-21T06:47:44.830 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout enabled 2026-03-21T06:47:44.833 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout active 2026-03-21T06:47:44.833 INFO:teuthology.orchestra.run.vm02.stdout:Unit ntp.service is enabled and running 2026-03-21T06:47:44.833 INFO:teuthology.orchestra.run.vm02.stdout:Host looks OK 2026-03-21T06:47:44.833 INFO:teuthology.orchestra.run.vm02.stdout:Cluster fsid: b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:44.833 INFO:teuthology.orchestra.run.vm02.stdout:Acquiring lock 140600730265936 on /run/cephadm/b16ecafc-24f1-11f1-8ede-8330751617ee.lock 2026-03-21T06:47:44.833 INFO:teuthology.orchestra.run.vm02.stdout:Lock 140600730265936 acquired on /run/cephadm/b16ecafc-24f1-11f1-8ede-8330751617ee.lock 2026-03-21T06:47:44.833 INFO:teuthology.orchestra.run.vm02.stdout:Verifying IP 192.168.123.102 port 3300 ... 2026-03-21T06:47:44.833 INFO:teuthology.orchestra.run.vm02.stdout:Verifying IP 192.168.123.102 port 6789 ... 2026-03-21T06:47:44.833 INFO:teuthology.orchestra.run.vm02.stdout:Base mon IP(s) is [192.168.123.102:3300, 192.168.123.102:6789], mon addrv is [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-21T06:47:44.835 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout default via 192.168.123.1 dev ens3 proto dhcp src 192.168.123.102 metric 100 2026-03-21T06:47:44.835 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 172.17.0.0/16 dev docker0 proto kernel scope link src 172.17.0.1 linkdown 2026-03-21T06:47:44.835 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 192.168.123.0/24 dev ens3 proto kernel scope link src 192.168.123.102 metric 100 2026-03-21T06:47:44.835 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 192.168.123.1 dev ens3 proto dhcp scope link src 192.168.123.102 metric 100 2026-03-21T06:47:44.836 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout ::1 dev lo proto kernel metric 256 pref medium 2026-03-21T06:47:44.836 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout fe80::/64 dev ens3 proto kernel metric 256 pref medium 2026-03-21T06:47:44.837 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 1: lo: mtu 65536 state UNKNOWN qlen 1000 2026-03-21T06:47:44.838 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout inet6 ::1/128 scope host 2026-03-21T06:47:44.838 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-21T06:47:44.838 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout 2: ens3: mtu 1500 state UP qlen 1000 2026-03-21T06:47:44.838 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout inet6 fe80::5055:ff:fe00:2/64 scope link 2026-03-21T06:47:44.838 INFO:teuthology.orchestra.run.vm02.stdout:/usr/sbin/ip: stdout valid_lft forever preferred_lft forever 2026-03-21T06:47:44.838 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.0/24` 2026-03-21T06:47:44.838 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.0/24` 2026-03-21T06:47:44.838 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.1/32` 2026-03-21T06:47:44.838 INFO:teuthology.orchestra.run.vm02.stdout:Mon IP `192.168.123.102` is in CIDR network `192.168.123.1/32` 2026-03-21T06:47:44.838 INFO:teuthology.orchestra.run.vm02.stdout:Inferred mon public CIDR from local network configuration ['192.168.123.0/24', '192.168.123.0/24', '192.168.123.1/32', '192.168.123.1/32'] 2026-03-21T06:47:44.839 INFO:teuthology.orchestra.run.vm02.stdout:Internal network (--cluster-network) has not been provided, OSD replication will default to the public_network 2026-03-21T06:47:44.839 INFO:teuthology.orchestra.run.vm02.stdout:Pulling container image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388... 2026-03-21T06:47:45.925 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout 70f8415b300f041766fa27faf7d5472699e32388: Pulling from ceph-ci/ceph 2026-03-21T06:47:45.925 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout Digest: sha256:5fe4cfc092321d85e013fc971c0a15337be7ad4a3c8cc47ec11af07596687bd6 2026-03-21T06:47:45.925 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout Status: Image is up to date for quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 2026-03-21T06:47:45.925 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/docker: stdout quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 2026-03-21T06:47:46.097 INFO:teuthology.orchestra.run.vm02.stdout:ceph: stdout ceph version 20.2.0-712-g70f8415b (70f8415b300f041766fa27faf7d5472699e32388) tentacle (stable) 2026-03-21T06:47:46.098 INFO:teuthology.orchestra.run.vm02.stdout:Ceph version: ceph version 20.2.0-712-g70f8415b (70f8415b300f041766fa27faf7d5472699e32388) tentacle (stable) 2026-03-21T06:47:46.098 INFO:teuthology.orchestra.run.vm02.stdout:Extracting ceph user uid/gid from container image... 2026-03-21T06:47:46.191 INFO:teuthology.orchestra.run.vm02.stdout:stat: stdout 167 167 2026-03-21T06:47:46.191 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial keys... 2026-03-21T06:47:46.326 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQASP75pFSYSERAAe8AaryfoWbiHRgwzA2YT1A== 2026-03-21T06:47:46.460 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQASP75pfy4rGRAANM9lHmw1xj4sQU5i4o0dqg== 2026-03-21T06:47:46.590 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-authtool: stdout AQASP75p803eIBAASEUGLfmNN7o6Sfqx9g9Mbw== 2026-03-21T06:47:46.590 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial monmap... 2026-03-21T06:47:46.730 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-21T06:47:46.730 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout setting min_mon_release = tentacle 2026-03-21T06:47:46.730 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: set fsid to b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:46.730 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: stdout /usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-21T06:47:46.730 INFO:teuthology.orchestra.run.vm02.stdout:monmaptool for a [v2:192.168.123.102:3300,v1:192.168.123.102:6789] on /usr/bin/monmaptool: monmap file /tmp/monmap 2026-03-21T06:47:46.730 INFO:teuthology.orchestra.run.vm02.stdout:setting min_mon_release = tentacle 2026-03-21T06:47:46.730 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: set fsid to b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:46.730 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/monmaptool: writing epoch 0 to /tmp/monmap (1 monitors) 2026-03-21T06:47:46.730 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:47:46.730 INFO:teuthology.orchestra.run.vm02.stdout:Creating mon... 2026-03-21T06:47:46.896 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.830+0000 7f8953d07d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-21T06:47:46.896 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.830+0000 7f8953d07d80 1 imported monmap: 2026-03-21T06:47:46.896 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr epoch 0 2026-03-21T06:47:46.896 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr last_changed 2026-03-21T06:47:46.690400+0000 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr min_mon_release 20 (tentacle) 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr election_strategy: 1 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.830+0000 7f8953d07d80 0 /usr/bin/ceph-mon: set fsid to b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Git sha 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Compile date 2026-03-13 17:47:44 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: DB SUMMARY 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: DB Session ID: DEK0I0SL5W4CMD3BYM1E 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 0, files: 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.error_if_exists: 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.create_if_missing: 1 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.env: 0x55e2e2136100 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.info_log: 0x55e2eb5e7ac0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.statistics: (nil) 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.use_fsync: 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.db_log_dir: 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.wal_dir: 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-21T06:47:46.897 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.write_buffer_manager: 0x55e2eb5eabe0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.unordered_write: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.row_cache: None 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.wal_filter: None 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.two_write_queues: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.wal_compression: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.atomic_flush: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.max_open_files: -1 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-21T06:47:46.898 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Compression algorithms supported: 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: kZSTD supported: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: kXpressCompression supported: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: kZlibCompression supported: 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.834+0000 7f8953d07d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: [db/db_impl/db_impl_open.cc:317] Creating manifest 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.merge_operator: 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_filter: None 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x55e2eb5e7580) 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks: 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr cache_index_and_filter_blocks_with_high_priority: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr pin_top_level_index_and_filter: 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr index_type: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr data_block_index_type: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr index_shortening: 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr data_block_hash_table_util_ratio: 0.750000 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr checksum: 4 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr no_block_cache: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache: 0x55e2eb607b90 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache_name: BinnedLRUCache 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache_options: 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr capacity : 536870912 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr num_shard_bits : 4 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr strict_capacity_limit : 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr high_pri_pool_ratio: 0.000 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_cache_compressed: (nil) 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr persistent_cache: (nil) 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_size: 4096 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_size_deviation: 10 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_restart_interval: 16 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr index_block_restart_interval: 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr metadata_block_size: 4096 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr partition_filters: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr use_delta_encoding: 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr filter_policy: bloomfilter 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr whole_key_filtering: 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr verify_compression: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr read_amp_bytes_per_bit: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr format_version: 5 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr enable_index_compression: 1 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr block_align: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr max_auto_readahead_size: 262144 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr prepopulate_block_cache: 0 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr initial_auto_readahead_size: 8192 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr num_file_reads_for_auto_readahead: 2 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-21T06:47:46.899 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compression: NoCompression 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.num_levels: 7 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.bloom_locality: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-21T06:47:46.900 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.ttl: 2592000 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.enable_blob_files: false 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.min_blob_size: 0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.838+0000 7f8953d07d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.842+0000 7f8953d07d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000001 succeeded,manifest_file_number is 1, next_file_number is 3, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.842+0000 7f8953d07d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.842+0000 7f8953d07d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: c5a495c2-60d1-43fe-a510-f93913a43d2b 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.842+0000 7f8953d07d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 5 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.842+0000 7f8953d07d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x55e2eb60a700 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.842+0000 7f8953d07d80 4 rocksdb: DB pointer 0x55e2eb658000 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.842+0000 7f894b491640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.842+0000 7f894b491640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** DB Stats ** 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Sum 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.00 0.00 0 0.000 0 0 0.0 0.0 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.901 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** Compaction Stats [default] ** 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Uptime(secs): 0.0 total, 0.0 interval 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Flush(GB): cumulative 0.000, interval 0.000 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(GB): cumulative 0.000, interval 0.000 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(Total Files): cumulative 0, interval 0 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(L0 Files): cumulative 0, interval 0 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr AddFile(Keys): cumulative 0, interval 0 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Cumulative compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Interval compaction: 0.00 GB write, 0.00 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Block cache BinnedLRUCache@0x55e2eb607b90#7 capacity: 512.00 MB usage: 0.00 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1.2e-05 secs_since: 0 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr Block cache entry stats(count,size,portion): Misc(1,0.00 KB,0%) 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr ** File Read Latency Histogram By Level [default] ** 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.846+0000 7f8953d07d80 4 rocksdb: [db/db_impl/db_impl.cc:496] Shutdown: canceling all background work 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.846+0000 7f8953d07d80 4 rocksdb: [db/db_impl/db_impl.cc:704] Shutdown complete 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph-mon: stderr debug 2026-03-21T06:47:46.846+0000 7f8953d07d80 0 /usr/bin/ceph-mon: created monfs at /var/lib/ceph/mon/ceph-a for mon.a 2026-03-21T06:47:46.902 INFO:teuthology.orchestra.run.vm02.stdout:create mon.a on 2026-03-21T06:47:47.260 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph.target → /etc/systemd/system/ceph.target. 2026-03-21T06:47:47.424 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/multi-user.target.wants/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee.target → /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee.target. 2026-03-21T06:47:47.425 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph.target.wants/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee.target → /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee.target. 2026-03-21T06:47:47.599 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.a 2026-03-21T06:47:47.599 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to reset failed state of unit ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.a.service: Unit ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.a.service not loaded. 2026-03-21T06:47:47.779 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee.target.wants/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.a.service → /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service. 2026-03-21T06:47:47.787 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-21T06:47:47.787 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to enable service . firewalld.service is not available 2026-03-21T06:47:47.787 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mon to start... 2026-03-21T06:47:47.787 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mon... 2026-03-21T06:47:48.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:47 vm02 bash[17177]: cluster 2026-03-21T06:47:47.919314+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout cluster: 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout id: b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout health: HEALTH_OK 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout services: 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon: 1 daemons, quorum a (age 0.233575s) [leader: a] 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr: no daemons active 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd: 0 osds: 0 up, 0 in 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout data: 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout pools: 0 pools, 0 pgs 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout objects: 0 objects, 0 B 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout usage: 0 B used, 0 B / 0 B avail 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout pgs: 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:mon is available 2026-03-21T06:47:48.198 INFO:teuthology.orchestra.run.vm02.stdout:Assimilating anything we can from ceph.conf... 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [global] 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout fsid = b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [osd] 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-21T06:47:48.446 INFO:teuthology.orchestra.run.vm02.stdout:Generating new minimal ceph.conf... 2026-03-21T06:47:48.635 INFO:teuthology.orchestra.run.vm02.stdout:Restarting the monitor... 2026-03-21T06:47:48.739 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 systemd[1]: Stopping Ceph mon.a for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T06:47:48.739 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17177]: debug 2026-03-21T06:47:48.662+0000 7f41be387640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-21T06:47:48.739 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17177]: debug 2026-03-21T06:47:48.666+0000 7f41be387640 -1 mon.a@0(leader) e1 *** Got Signal Terminated *** 2026-03-21T06:47:48.788 INFO:teuthology.orchestra.run.vm02.stdout:Setting public_network to 192.168.123.0/24,192.168.123.1/32 in global config section 2026-03-21T06:47:49.006 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17563]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-mon-a 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.a.service: Deactivated successfully. 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 systemd[1]: Stopped Ceph mon.a for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 systemd[1]: Started Ceph mon.a for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.894+0000 7fe932690d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.894+0000 7fe932690d80 0 ceph version 20.2.0-712-g70f8415b (70f8415b300f041766fa27faf7d5472699e32388) tentacle (stable - Debug), process ceph-mon, pid 7 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.894+0000 7fe932690d80 0 pidfile_write: ignore empty --pid-file 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 0 load: jerasure load: lrc 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Git sha 0 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Compile date 2026-03-13 17:47:44 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: DB SUMMARY 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: DB Session ID: BWT82ZTZYTR8L2XOV2TN 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: CURRENT file: CURRENT 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: MANIFEST file: MANIFEST-000010 size: 179 Bytes 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-a/store.db dir, Total Num: 1, files: 000008.sst 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-a/store.db: 000009.log size: 77013 ; 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.error_if_exists: 0 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.create_if_missing: 0 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.env: 0x563ad8ba6100 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.info_log: 0x563b01665b00 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.statistics: (nil) 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.use_fsync: 0 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-21T06:47:49.007 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.db_log_dir: 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.wal_dir: 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.write_buffer_manager: 0x563b01668500 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.unordered_write: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.row_cache: None 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.wal_filter: None 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.two_write_queues: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.wal_compression: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.atomic_flush: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-21T06:47:49.008 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_open_files: -1 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Compression algorithms supported: 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: kZSTD supported: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: kXpressCompression supported: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: kZlibCompression supported: 1 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.merge_operator: 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_filter: None 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x563b01664240) 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cache_index_and_filter_blocks: 1 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: pin_top_level_index_and_filter: 1 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: index_type: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: data_block_index_type: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: index_shortening: 1 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: data_block_hash_table_util_ratio: 0.750000 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: checksum: 4 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: no_block_cache: 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: block_cache: 0x563b01685b90 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: block_cache_name: BinnedLRUCache 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: block_cache_options: 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: capacity : 536870912 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: num_shard_bits : 4 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: strict_capacity_limit : 0 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: high_pri_pool_ratio: 0.000 2026-03-21T06:47:49.009 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: block_cache_compressed: (nil) 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: persistent_cache: (nil) 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: block_size: 4096 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: block_size_deviation: 10 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: block_restart_interval: 16 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: index_block_restart_interval: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: metadata_block_size: 4096 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: partition_filters: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: use_delta_encoding: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: filter_policy: bloomfilter 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: whole_key_filtering: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: verify_compression: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: read_amp_bytes_per_bit: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: format_version: 5 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: enable_index_compression: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: block_align: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: max_auto_readahead_size: 262144 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: prepopulate_block_cache: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: initial_auto_readahead_size: 8192 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: num_file_reads_for_auto_readahead: 2 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compression: NoCompression 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.num_levels: 7 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-21T06:47:49.010 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.bloom_locality: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.ttl: 2592000 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.enable_blob_files: false 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.min_blob_size: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.898+0000 7fe932690d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.902+0000 7fe932690d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-a/store.db/MANIFEST-000010 succeeded,manifest_file_number is 10, next_file_number is 12, last_sequence is 5, log_number is 5,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 5 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.902+0000 7fe932690d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 5 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.902+0000 7fe932690d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: c5a495c2-60d1-43fe-a510-f93913a43d2b 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.902+0000 7fe932690d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774075668906300, "job": 1, "event": "recovery_started", "wal_files": [9]} 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.902+0000 7fe932690d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #9 mode 2 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.906+0000 7fe932690d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774075668908054, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 13, "file_size": 73999, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 8, "largest_seqno": 231, "table_properties": {"data_size": 72278, "index_size": 174, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 517, "raw_key_size": 9926, "raw_average_key_size": 49, "raw_value_size": 66661, "raw_average_value_size": 330, "num_data_blocks": 8, "num_entries": 202, "num_filter_entries": 202, "num_deletions": 3, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1774075668, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "c5a495c2-60d1-43fe-a510-f93913a43d2b", "db_session_id": "BWT82ZTZYTR8L2XOV2TN", "orig_file_number": 13, "seqno_to_time_mapping": "N/A"}} 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.906+0000 7fe932690d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774075668908164, "job": 1, "event": "recovery_finished"} 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.906+0000 7fe932690d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 15 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.906+0000 7fe932690d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-a/store.db/000009.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.906+0000 7fe932690d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x563b01688700 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.906+0000 7fe932690d80 4 rocksdb: DB pointer 0x563b016d2000 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.906+0000 7fe928419640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.906+0000 7fe928419640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: ** DB Stats ** 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: ** Compaction Stats [default] ** 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: L0 2/0 74.12 KB 0.5 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 49.9 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Sum 2/0 74.12 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 49.9 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 49.9 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: ** Compaction Stats [default] ** 2026-03-21T06:47:49.011 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 49.9 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: AddFile(Total Files): cumulative 0, interval 0 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: AddFile(Keys): cumulative 0, interval 0 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Cumulative compaction: 0.00 GB write, 6.92 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Interval compaction: 0.00 GB write, 6.92 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Block cache BinnedLRUCache@0x563b01685b90#7 capacity: 512.00 MB usage: 1.08 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 1e-05 secs_since: 0 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: Block cache entry stats(count,size,portion): FilterBlock(2,0.72 KB,0.000137091%) IndexBlock(2,0.36 KB,6.85453e-05%) Misc(1,0.00 KB,0%) 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: ** File Read Latency Histogram By Level [default] ** 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.914+0000 7fe932690d80 0 starting mon.a rank 0 at public addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] at bind addrs [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon_data /var/lib/ceph/mon/ceph-a fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.914+0000 7fe932690d80 1 mon.a@-1(???) e1 preinit fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.914+0000 7fe932690d80 0 mon.a@-1(???).mds e1 new map 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.914+0000 7fe932690d80 0 mon.a@-1(???).mds e1 print_map 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: e1 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: btime 2026-03-21T06:47:47:923913+0000 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: legacy client fscid: -1 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: No filesystems configured 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.914+0000 7fe932690d80 0 mon.a@-1(???).osd e1 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.914+0000 7fe932690d80 0 mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.914+0000 7fe932690d80 0 mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.914+0000 7fe932690d80 0 mon.a@-1(???).osd e1 crush map has features 288514050185494528, adjusting msgr requires 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: debug 2026-03-21T06:47:48.914+0000 7fe932690d80 1 mon.a@-1(???).paxosservice(auth 1..2) refresh upgraded, format 0 -> 3 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920244+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920244+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920278+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920278+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920283+0000 mon.a (mon.0) 3 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920283+0000 mon.a (mon.0) 3 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920287+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-21T06:47:46.690400+0000 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920287+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-21T06:47:46.690400+0000 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920297+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920297+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920302+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920302+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920311+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920311+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920314+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920314+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920633+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920633+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920648+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.920648+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.921256+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-21T06:47:49.012 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:48 vm02 bash[17657]: cluster 2026-03-21T06:47:48.921256+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-21T06:47:49.034 INFO:teuthology.orchestra.run.vm02.stdout:Wrote config to /etc/ceph/ceph.conf 2026-03-21T06:47:49.034 INFO:teuthology.orchestra.run.vm02.stdout:Wrote keyring to /etc/ceph/ceph.client.admin.keyring 2026-03-21T06:47:49.034 INFO:teuthology.orchestra.run.vm02.stdout:Creating mgr... 2026-03-21T06:47:49.034 INFO:teuthology.orchestra.run.vm02.stdout:Verifying port 0.0.0.0:9283 ... 2026-03-21T06:47:49.035 INFO:teuthology.orchestra.run.vm02.stdout:Verifying port 0.0.0.0:8765 ... 2026-03-21T06:47:49.225 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 1 from systemctl reset-failed ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mgr.x 2026-03-21T06:47:49.225 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Failed to reset failed state of unit ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mgr.x.service: Unit ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mgr.x.service not loaded. 2026-03-21T06:47:49.331 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:49 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:47:49.397 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stderr Created symlink /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee.target.wants/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mgr.x.service → /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service. 2026-03-21T06:47:49.404 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-21T06:47:49.404 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to enable service . firewalld.service is not available 2026-03-21T06:47:49.404 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-21T06:47:49.404 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to open ports <[9283, 8765]>. firewalld.service is not available 2026-03-21T06:47:49.404 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr to start... 2026-03-21T06:47:49.404 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr... 2026-03-21T06:47:49.601 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:49 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "b16ecafc-24f1-11f1-8ede-8330751617ee", 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-21T06:47:49.630 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "tentacle", 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-21T06:47:47:923913+0000", 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-21T06:47:49.631 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs" 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-21T06:47:47.924677+0000", 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-21T06:47:49.632 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (1/15)... 2026-03-21T06:47:50.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:49 vm02 bash[17657]: audit 2026-03-21T06:47:48.990782+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1389498430' entity='client.admin' 2026-03-21T06:47:50.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:49 vm02 bash[17657]: audit 2026-03-21T06:47:48.990782+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1389498430' entity='client.admin' 2026-03-21T06:47:50.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:49 vm02 bash[17657]: audit 2026-03-21T06:47:49.586159+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/4046014269' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:47:50.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:49 vm02 bash[17657]: audit 2026-03-21T06:47:49.586159+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/4046014269' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:47:51.146 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:47:50 vm02 bash[17927]: from numpy import show_config as show_numpy_config 2026-03-21T06:47:51.861 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "b16ecafc-24f1-11f1-8ede-8330751617ee", 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 2, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "tentacle", 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-21T06:47:51.862 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-21T06:47:47:923913+0000", 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": false, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs" 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-21T06:47:47.924677+0000", 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-21T06:47:51.863 INFO:teuthology.orchestra.run.vm02.stdout:mgr not available, waiting (2/15)... 2026-03-21T06:47:52.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:51 vm02 bash[17657]: audit 2026-03-21T06:47:51.810831+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/788823380' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:47:52.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:51 vm02 bash[17657]: audit 2026-03-21T06:47:51.810831+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/788823380' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: cluster 2026-03-21T06:47:52.480256+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon x 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: cluster 2026-03-21T06:47:52.480256+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon x 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: cluster 2026-03-21T06:47:52.484714+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: x(active, starting, since 0.00461082s) 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: cluster 2026-03-21T06:47:52.484714+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: x(active, starting, since 0.00461082s) 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.486227+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.486227+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.486600+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.486600+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.486956+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.486956+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.487347+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:47:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.487347+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.487688+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.487688+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: cluster 2026-03-21T06:47:52.494572+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon x is now available 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: cluster 2026-03-21T06:47:52.494572+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon x is now available 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.503203+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.503203+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.507496+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.507496+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.507813+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.507813+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.511606+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.511606+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.514731+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:47:52.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:52 vm02 bash[17657]: audit 2026-03-21T06:47:52.514731+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:47:54.193 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:54.193 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsid": "b16ecafc-24f1-11f1-8ede-8330751617ee", 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "health": { 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "status": "HEALTH_OK", 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "checks": {}, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mutes": [] 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "election_epoch": 5, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum": [ 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 0 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_names": [ 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "a" 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "quorum_age": 5, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "monmap": { 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "min_mon_release_name": "tentacle", 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_mons": 1 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osdmap": { 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_osds": 0, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_up_osds": 0, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_up_since": 0, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_in_osds": 0, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "osd_in_since": 0, 2026-03-21T06:47:54.194 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_remapped_pgs": 0 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgmap": { 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "pgs_by_state": [], 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pgs": 0, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_pools": 0, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_objects": 0, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "data_bytes": 0, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_used": 0, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_avail": 0, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "bytes_total": 0 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "fsmap": { 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "btime": "2026-03-21T06:47:47:923913+0000", 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "by_rank": [], 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "up:standby": 0 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap": { 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standbys": 0, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modules": [ 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "iostat", 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "nfs" 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ], 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "servicemap": { 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 1, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "modified": "2026-03-21T06:47:47.924677+0000", 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "services": {} 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout }, 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "progress_events": {} 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-21T06:47:54.195 INFO:teuthology.orchestra.run.vm02.stdout:mgr is available 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [global] 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout fsid = b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_cluster_log_file_level = debug 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_host = [v2:192.168.123.102:3300,v1:192.168.123.102:6789] 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_pg_remap = true 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_osd_allow_primary_affinity = true 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mon_warn_on_no_sortbitwise = false 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_crush_chooseleaf_type = 0 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [mgr] 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout mgr/telemetry/nag = false 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout [osd] 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_map_max_advance = 10 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout osd_sloppy_crc = true 2026-03-21T06:47:54.484 INFO:teuthology.orchestra.run.vm02.stdout:Enabling cephadm module... 2026-03-21T06:47:54.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:54 vm02 bash[17657]: cluster 2026-03-21T06:47:53.490150+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: x(active, since 1.01005s) 2026-03-21T06:47:54.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:54 vm02 bash[17657]: cluster 2026-03-21T06:47:53.490150+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: x(active, since 1.01005s) 2026-03-21T06:47:54.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:54 vm02 bash[17657]: audit 2026-03-21T06:47:54.153396+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.102:0/1916943896' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:47:54.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:54 vm02 bash[17657]: audit 2026-03-21T06:47:54.153396+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.102:0/1916943896' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:47:54.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:54 vm02 bash[17657]: audit 2026-03-21T06:47:54.442282+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.102:0/1165510801' entity='client.admin' cmd={"prefix": "config assimilate-conf"} : dispatch 2026-03-21T06:47:54.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:54 vm02 bash[17657]: audit 2026-03-21T06:47:54.442282+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.102:0/1165510801' entity='client.admin' cmd={"prefix": "config assimilate-conf"} : dispatch 2026-03-21T06:47:55.854 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:47:55 vm02 bash[17927]: ignoring --setuser ceph since I am not root 2026-03-21T06:47:55.854 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:47:55 vm02 bash[17927]: ignoring --setgroup ceph since I am not root 2026-03-21T06:47:55.854 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:55 vm02 bash[17657]: cluster 2026-03-21T06:47:54.494219+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e4: x(active, since 2s) 2026-03-21T06:47:55.854 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:55 vm02 bash[17657]: cluster 2026-03-21T06:47:54.494219+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e4: x(active, since 2s) 2026-03-21T06:47:55.855 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:55 vm02 bash[17657]: audit 2026-03-21T06:47:54.801057+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "cephadm"} : dispatch 2026-03-21T06:47:55.855 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:55 vm02 bash[17657]: audit 2026-03-21T06:47:54.801057+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "cephadm"} : dispatch 2026-03-21T06:47:55.883 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-21T06:47:55.883 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 5, 2026-03-21T06:47:55.883 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-21T06:47:55.883 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "active_name": "x", 2026-03-21T06:47:55.883 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-21T06:47:55.883 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-21T06:47:55.883 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for the mgr to restart... 2026-03-21T06:47:55.883 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr epoch 5... 2026-03-21T06:47:56.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:56 vm02 bash[17657]: audit 2026-03-21T06:47:55.502031+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-21T06:47:56.771 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:56 vm02 bash[17657]: audit 2026-03-21T06:47:55.502031+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-21T06:47:56.771 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:56 vm02 bash[17657]: cluster 2026-03-21T06:47:55.504242+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e5: x(active, since 3s) 2026-03-21T06:47:56.771 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:56 vm02 bash[17657]: cluster 2026-03-21T06:47:55.504242+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e5: x(active, since 3s) 2026-03-21T06:47:56.771 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:56 vm02 bash[17657]: audit 2026-03-21T06:47:55.836973+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.102:0/4113217554' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:47:56.771 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:56 vm02 bash[17657]: audit 2026-03-21T06:47:55.836973+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.102:0/4113217554' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:47:57.146 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:47:56 vm02 bash[17927]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-21T06:47:57.146 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:47:56 vm02 bash[17927]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-21T06:47:57.146 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:47:56 vm02 bash[17927]: from numpy import show_config as show_numpy_config 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: cluster 2026-03-21T06:47:58.527847+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: cluster 2026-03-21T06:47:58.527847+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: cluster 2026-03-21T06:47:58.528170+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon x 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: cluster 2026-03-21T06:47:58.528170+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon x 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: cluster 2026-03-21T06:47:58.533539+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: cluster 2026-03-21T06:47:58.533539+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: cluster 2026-03-21T06:47:58.533656+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e6: x(active, starting, since 0.00561459s) 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: cluster 2026-03-21T06:47:58.533656+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e6: x(active, starting, since 0.00561459s) 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: audit 2026-03-21T06:47:58.535351+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: audit 2026-03-21T06:47:58.535351+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: audit 2026-03-21T06:47:58.536553+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: audit 2026-03-21T06:47:58.536553+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: audit 2026-03-21T06:47:58.537630+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: audit 2026-03-21T06:47:58.537630+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: audit 2026-03-21T06:47:58.537826+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: audit 2026-03-21T06:47:58.537826+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: audit 2026-03-21T06:47:58.537993+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: audit 2026-03-21T06:47:58.537993+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: cluster 2026-03-21T06:47:58.546117+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon x is now available 2026-03-21T06:47:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:47:58 vm02 bash[17657]: cluster 2026-03-21T06:47:58.546117+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon x is now available 2026-03-21T06:47:59.611 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-21T06:47:59.611 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 7, 2026-03-21T06:47:59.611 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-21T06:47:59.611 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-21T06:47:59.611 INFO:teuthology.orchestra.run.vm02.stdout:mgr epoch 5 is available 2026-03-21T06:47:59.611 INFO:teuthology.orchestra.run.vm02.stdout:Verifying orchestrator module is enabled... 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.047042+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.047042+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.051356+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.051356+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: cephadm 2026-03-21T06:47:59.051977+0000 mgr.x (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: cephadm 2026-03-21T06:47:59.051977+0000 mgr.x (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.054419+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.054419+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.060887+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.060887+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.064663+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.064663+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.067675+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.067675+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.072257+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.072257+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.074408+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.074408+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: cluster 2026-03-21T06:47:59.537842+0000 mon.a (mon.0) 54 : cluster [DBG] mgrmap e7: x(active, since 1.0098s) 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: cluster 2026-03-21T06:47:59.537842+0000 mon.a (mon.0) 54 : cluster [DBG] mgrmap e7: x(active, since 1.0098s) 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.538814+0000 mgr.x (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.538814+0000 mgr.x (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.543003+0000 mgr.x (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.543003+0000 mgr.x (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.908263+0000 mon.a (mon.0) 55 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "orchestrator"} : dispatch 2026-03-21T06:48:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:00 vm02 bash[17657]: audit 2026-03-21T06:47:59.908263+0000 mon.a (mon.0) 55 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "orchestrator"} : dispatch 2026-03-21T06:48:00.602 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stderr module 'orchestrator' is already enabled (always-on) 2026-03-21T06:48:00.602 INFO:teuthology.orchestra.run.vm02.stdout:Setting orchestrator backend to cephadm... 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cephadm 2026-03-21T06:48:00.100180+0000 mgr.x (mgr.14118) 4 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTING 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cephadm 2026-03-21T06:48:00.100180+0000 mgr.x (mgr.14118) 4 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTING 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cephadm 2026-03-21T06:48:00.201223+0000 mgr.x (mgr.14118) 5 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cephadm 2026-03-21T06:48:00.201223+0000 mgr.x (mgr.14118) 5 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cephadm 2026-03-21T06:48:00.308850+0000 mgr.x (mgr.14118) 6 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cephadm 2026-03-21T06:48:00.308850+0000 mgr.x (mgr.14118) 6 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cephadm 2026-03-21T06:48:00.308904+0000 mgr.x (mgr.14118) 7 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTED 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cephadm 2026-03-21T06:48:00.308904+0000 mgr.x (mgr.14118) 7 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTED 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cephadm 2026-03-21T06:48:00.309376+0000 mgr.x (mgr.14118) 8 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Client ('192.168.123.102', 32874) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cephadm 2026-03-21T06:48:00.309376+0000 mgr.x (mgr.14118) 8 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Client ('192.168.123.102', 32874) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: audit 2026-03-21T06:48:00.309379+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: audit 2026-03-21T06:48:00.309379+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: audit 2026-03-21T06:48:00.537516+0000 mon.a (mon.0) 57 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "orchestrator"}]': finished 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: audit 2026-03-21T06:48:00.537516+0000 mon.a (mon.0) 57 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "orchestrator"}]': finished 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cluster 2026-03-21T06:48:00.539788+0000 mon.a (mon.0) 58 : cluster [DBG] mgrmap e8: x(active, since 2s) 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: cluster 2026-03-21T06:48:00.539788+0000 mon.a (mon.0) 58 : cluster [DBG] mgrmap e8: x(active, since 2s) 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: audit 2026-03-21T06:48:00.866761+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: audit 2026-03-21T06:48:00.866761+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: audit 2026-03-21T06:48:00.871909+0000 mon.a (mon.0) 60 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:01 vm02 bash[17657]: audit 2026-03-21T06:48:00.871909+0000 mon.a (mon.0) 60 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:01.226 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout value unchanged 2026-03-21T06:48:01.226 INFO:teuthology.orchestra.run.vm02.stdout:Generating ssh key... 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: Generating public/private ed25519 key pair. 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: Your identification has been saved in /tmp/tmp7vgusnzk/key 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: Your public key has been saved in /tmp/tmp7vgusnzk/key.pub 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: The key fingerprint is: 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: SHA256:EYfIA3whQl6rc/ezzyjONnSNGeeRRTdtwfeelDnkO1I ceph-b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: The key's randomart image is: 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: +--[ED25519 256]--+ 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: | .o.+o.o..... +o.| 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: | . o.o= .o .. ++| 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: | . .. .. o o.=| 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: | . ..+ E.| 2026-03-21T06:48:01.777 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: | o . . SB . o.+| 2026-03-21T06:48:01.778 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: | o ...+ o . +.| 2026-03-21T06:48:01.778 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: | . .o . .| 2026-03-21T06:48:01.778 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: | .+ = | 2026-03-21T06:48:01.778 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: | oooo.o | 2026-03-21T06:48:01.778 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:01 vm02 bash[17927]: +----[SHA256]-----+ 2026-03-21T06:48:01.801 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDqVQU8deI5GkrkKLViWR7HEg9MGwqno6Oh6IzEbn/fy ceph-b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:01.801 INFO:teuthology.orchestra.run.vm02.stdout:Wrote public SSH key to /home/ubuntu/cephtest/ceph.pub 2026-03-21T06:48:01.801 INFO:teuthology.orchestra.run.vm02.stdout:Adding key to root@localhost authorized_keys... 2026-03-21T06:48:01.802 INFO:teuthology.orchestra.run.vm02.stdout:Adding host vm02... 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: audit 2026-03-21T06:48:00.862870+0000 mgr.x (mgr.14118) 9 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: audit 2026-03-21T06:48:00.862870+0000 mgr.x (mgr.14118) 9 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: audit 2026-03-21T06:48:01.184829+0000 mgr.x (mgr.14118) 10 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: audit 2026-03-21T06:48:01.184829+0000 mgr.x (mgr.14118) 10 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: audit 2026-03-21T06:48:01.460422+0000 mgr.x (mgr.14118) 11 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: audit 2026-03-21T06:48:01.460422+0000 mgr.x (mgr.14118) 11 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: cephadm 2026-03-21T06:48:01.460668+0000 mgr.x (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: cephadm 2026-03-21T06:48:01.460668+0000 mgr.x (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: audit 2026-03-21T06:48:01.477702+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: audit 2026-03-21T06:48:01.477702+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: audit 2026-03-21T06:48:01.480261+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:02.629 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:02 vm02 bash[17657]: audit 2026-03-21T06:48:01.480261+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:03.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:03 vm02 bash[17657]: audit 2026-03-21T06:48:01.760765+0000 mgr.x (mgr.14118) 13 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:03.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:03 vm02 bash[17657]: audit 2026-03-21T06:48:01.760765+0000 mgr.x (mgr.14118) 13 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:03.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:03 vm02 bash[17657]: audit 2026-03-21T06:48:02.049007+0000 mgr.x (mgr.14118) 14 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:03.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:03 vm02 bash[17657]: audit 2026-03-21T06:48:02.049007+0000 mgr.x (mgr.14118) 14 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:03.958 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Added host 'vm02' with addr '192.168.123.102' 2026-03-21T06:48:03.958 INFO:teuthology.orchestra.run.vm02.stdout:Deploying unmanaged mon service... 2026-03-21T06:48:04.290 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Scheduled mon update... 2026-03-21T06:48:04.291 INFO:teuthology.orchestra.run.vm02.stdout:Deploying unmanaged mgr service... 2026-03-21T06:48:04.588 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:04 vm02 bash[17657]: cephadm 2026-03-21T06:48:02.587117+0000 mgr.x (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-21T06:48:04.588 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:04 vm02 bash[17657]: cephadm 2026-03-21T06:48:02.587117+0000 mgr.x (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-21T06:48:04.588 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:04 vm02 bash[17657]: audit 2026-03-21T06:48:03.905211+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:04.588 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:04 vm02 bash[17657]: audit 2026-03-21T06:48:03.905211+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:04.588 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:04 vm02 bash[17657]: audit 2026-03-21T06:48:03.906295+0000 mon.a (mon.0) 64 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:04.588 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:04 vm02 bash[17657]: audit 2026-03-21T06:48:03.906295+0000 mon.a (mon.0) 64 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:04.588 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:04 vm02 bash[17657]: audit 2026-03-21T06:48:04.240830+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:04.588 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:04 vm02 bash[17657]: audit 2026-03-21T06:48:04.240830+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:04.643 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout Scheduled mgr update... 2026-03-21T06:48:05.324 INFO:teuthology.orchestra.run.vm02.stdout:Enabling the dashboard module... 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: cephadm 2026-03-21T06:48:03.906013+0000 mgr.x (mgr.14118) 16 : cephadm [INF] Added host vm02 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: cephadm 2026-03-21T06:48:03.906013+0000 mgr.x (mgr.14118) 16 : cephadm [INF] Added host vm02 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:04.235767+0000 mgr.x (mgr.14118) 17 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:04.235767+0000 mgr.x (mgr.14118) 17 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: cephadm 2026-03-21T06:48:04.236801+0000 mgr.x (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: cephadm 2026-03-21T06:48:04.236801+0000 mgr.x (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:04.562737+0000 mgr.x (mgr.14118) 19 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:04.562737+0000 mgr.x (mgr.14118) 19 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: cephadm 2026-03-21T06:48:04.563606+0000 mgr.x (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: cephadm 2026-03-21T06:48:04.563606+0000 mgr.x (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:04.567739+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:04.567739+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:04.919608+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.102:0/3282338898' entity='client.admin' 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:04.919608+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.102:0/3282338898' entity='client.admin' 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:05.241068+0000 mon.a (mon.0) 68 : audit [INF] from='client.? 192.168.123.102:0/705651056' entity='client.admin' 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:05.241068+0000 mon.a (mon.0) 68 : audit [INF] from='client.? 192.168.123.102:0/705651056' entity='client.admin' 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:05.399776+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:05 vm02 bash[17657]: audit 2026-03-21T06:48:05.399776+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:06 vm02 bash[17657]: audit 2026-03-21T06:48:05.656724+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "dashboard"} : dispatch 2026-03-21T06:48:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:06 vm02 bash[17657]: audit 2026-03-21T06:48:05.656724+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "dashboard"} : dispatch 2026-03-21T06:48:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:06 vm02 bash[17657]: audit 2026-03-21T06:48:05.737887+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:06 vm02 bash[17657]: audit 2026-03-21T06:48:05.737887+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:06.896 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:06 vm02 bash[17927]: ignoring --setuser ceph since I am not root 2026-03-21T06:48:06.896 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:06 vm02 bash[17927]: ignoring --setgroup ceph since I am not root 2026-03-21T06:48:06.988 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-21T06:48:06.989 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "epoch": 9, 2026-03-21T06:48:06.989 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "available": true, 2026-03-21T06:48:06.989 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "active_name": "x", 2026-03-21T06:48:06.989 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "num_standby": 0 2026-03-21T06:48:06.989 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-21T06:48:06.989 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for the mgr to restart... 2026-03-21T06:48:06.989 INFO:teuthology.orchestra.run.vm02.stdout:Waiting for mgr epoch 9... 2026-03-21T06:48:07.864 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:07 vm02 bash[17657]: audit 2026-03-21T06:48:06.573803+0000 mon.a (mon.0) 72 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-21T06:48:07.864 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:07 vm02 bash[17657]: audit 2026-03-21T06:48:06.573803+0000 mon.a (mon.0) 72 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-21T06:48:07.864 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:07 vm02 bash[17657]: cluster 2026-03-21T06:48:06.576279+0000 mon.a (mon.0) 73 : cluster [DBG] mgrmap e9: x(active, since 8s) 2026-03-21T06:48:07.864 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:07 vm02 bash[17657]: cluster 2026-03-21T06:48:06.576279+0000 mon.a (mon.0) 73 : cluster [DBG] mgrmap e9: x(active, since 8s) 2026-03-21T06:48:07.864 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:07 vm02 bash[17657]: audit 2026-03-21T06:48:06.947047+0000 mon.a (mon.0) 74 : audit [DBG] from='client.? 192.168.123.102:0/2257204228' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:48:07.864 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:07 vm02 bash[17657]: audit 2026-03-21T06:48:06.947047+0000 mon.a (mon.0) 74 : audit [DBG] from='client.? 192.168.123.102:0/2257204228' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:48:08.146 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:07 vm02 bash[17927]: /lib64/python3.9/site-packages/scipy/__init__.py:73: UserWarning: NumPy was imported from a Python sub-interpreter but NumPy does not properly support sub-interpreters. This will likely work for most users but might cause hard to track down issues or subtle bugs. A common user of the rare sub-interpreter feature is wsgi which also allows single-interpreter mode. 2026-03-21T06:48:08.146 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:07 vm02 bash[17927]: Improvements in the case of bugs are welcome, but is not on the NumPy roadmap, and full support may require significant effort to achieve. 2026-03-21T06:48:08.146 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:07 vm02 bash[17927]: from numpy import show_config as show_numpy_config 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: cluster 2026-03-21T06:48:09.528899+0000 mon.a (mon.0) 75 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: cluster 2026-03-21T06:48:09.528899+0000 mon.a (mon.0) 75 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: cluster 2026-03-21T06:48:09.529177+0000 mon.a (mon.0) 76 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: cluster 2026-03-21T06:48:09.529177+0000 mon.a (mon.0) 76 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: cluster 2026-03-21T06:48:09.534390+0000 mon.a (mon.0) 77 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: cluster 2026-03-21T06:48:09.534390+0000 mon.a (mon.0) 77 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: cluster 2026-03-21T06:48:09.534484+0000 mon.a (mon.0) 78 : cluster [DBG] mgrmap e10: x(active, starting, since 0.00543034s) 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: cluster 2026-03-21T06:48:09.534484+0000 mon.a (mon.0) 78 : cluster [DBG] mgrmap e10: x(active, starting, since 0.00543034s) 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: audit 2026-03-21T06:48:09.535036+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: audit 2026-03-21T06:48:09.535036+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: audit 2026-03-21T06:48:09.535833+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: audit 2026-03-21T06:48:09.535833+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: audit 2026-03-21T06:48:09.536634+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: audit 2026-03-21T06:48:09.536634+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: audit 2026-03-21T06:48:09.536754+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: audit 2026-03-21T06:48:09.536754+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: audit 2026-03-21T06:48:09.536847+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: audit 2026-03-21T06:48:09.536847+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: cluster 2026-03-21T06:48:09.542468+0000 mon.a (mon.0) 84 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:09.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:09 vm02 bash[17657]: cluster 2026-03-21T06:48:09.542468+0000 mon.a (mon.0) 84 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:10.588 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout { 2026-03-21T06:48:10.588 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "mgrmap_epoch": 11, 2026-03-21T06:48:10.588 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout "initialized": true 2026-03-21T06:48:10.588 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout } 2026-03-21T06:48:10.588 INFO:teuthology.orchestra.run.vm02.stdout:mgr epoch 9 is available 2026-03-21T06:48:10.588 INFO:teuthology.orchestra.run.vm02.stdout:Using certmgr to generate dashboard self-signed certificate... 2026-03-21T06:48:10.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:10 vm02 bash[17657]: audit 2026-03-21T06:48:09.733700+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:10.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:10 vm02 bash[17657]: audit 2026-03-21T06:48:09.733700+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:10.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:10 vm02 bash[17657]: audit 2026-03-21T06:48:09.747375+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:10.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:10 vm02 bash[17657]: audit 2026-03-21T06:48:09.747375+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:10.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:10 vm02 bash[17657]: audit 2026-03-21T06:48:09.754178+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:10.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:10 vm02 bash[17657]: audit 2026-03-21T06:48:09.754178+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:10.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:10 vm02 bash[17657]: cluster 2026-03-21T06:48:10.540431+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e11: x(active, since 1.01138s) 2026-03-21T06:48:10.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:10 vm02 bash[17657]: cluster 2026-03-21T06:48:10.540431+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e11: x(active, since 1.01138s) 2026-03-21T06:48:11.816 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout {"cert": "-----BEGIN CERTIFICATE-----\nMIIE+TCCAuGgAwIBAgIUD0abySiBeT+Ya5YL/9XQkiiesZkwDQYJKoZIhvcNAQEL\nBQAwFzEVMBMGA1UEAwwMY2VwaGFkbS1yb290MB4XDTI2MDMyMTA2NDgxMVoXDTI5\nMDMyMDA2NDgxMVowGjEYMBYGA1UEAwwPMTkyLjE2OC4xMjMuMTAyMIICIjANBgkq\nhkiG9w0BAQEFAAOCAg8AMIICCgKCAgEAwUdLigUqm4ihe1tN3HWP8iefRCv5BKBZ\nNP7e3i5WDh8Q0XLHDYOLEEu7RwfdGYll/Lr7xmJrpPINKZIJ/7Klg4GAcI34fNUC\nXlSUSZjQ/3fq/YWWbit1RbjoNjOWNvzzNBI4m53WCLR4LnTorcQx9NPnaolf3xwX\nMua9LHbWj3Jc4rnbVV7PKLSTyEY07WjdLhqnFRIYnNPjk1e9VRhzq9c8D360FTuE\nyICa9oLP6KGJxGPXfxvk45M4ztEmgdOmmunjRmxRmDXUEj5ZmGJHm75dwjYZsktD\neKJVO64+iuLzS9cnq5o4ERxpHkLnDAbIZ35YCLiJs8EGV4xygo7xVfYkReLooLHb\nCCVv9qsfdizi2t/4LhxxFpSUGnPaior+Lxic2miu+arZ+2E5EWXPz/QoP65JCfCW\nDxd7EbDLQ8bcEo3LwM6sgmeac184H3bPlfMG/RNn+qI71stdFpgCuM7hD5h+PChE\nBrhUnCn3fzYTBFHLUHwzedVBLKgk9LxecPKIa4vIfShUsD1XhktoaliiKqOq4wgG\n4z0iJ2wTsIpfJO4RTLsGF4pf+cbtpugVmCuMLnH4c4jb2br4D4FwdraGOhBgjoE5\nzoRdTDh4JOL46M28kaO8v4uSB2tVgOkY9aLaHt1urhu9gxNikwzLIvJIS8oRANhi\nu6Dpbi6c7BkCAwEAAaM6MDgwKAYDVR0RBCEwH4IEdm0wMoIRZGFzaGJvYXJkX3Nl\ncnZlcnOHBMCoe2YwDAYDVR0TAQH/BAIwADANBgkqhkiG9w0BAQsFAAOCAgEATV6d\nu0voFomMJ2swE9ItlKVNGpTbSG0UmsEnqc1bHfLfnFGQw/P5lddAY8d0wiSdjVtp\nTfBBITDS7Ik02zAtGsFJWNIt14h7BuhPLxBPKQh+8sx+hDACWBdH4an9CBuFEK9D\naUjYeLgHGulnWKw5yRmaWgYLLUDAyuRU/NXyn8poJ0cWzsKn01UfT4iU62D6n9Kj\n8Rmb+yfzWmxclY+U3HHYiwo1ryIAFHLeZs12wc99q1H6yD5EkQOpLd25uATifJfh\n/cMCnZEaPBLp64R3Fo7WUXnuZt9g7inJhKc/b7laPyTqXzzsFnDyldBerj85YOoU\nzFnlRRmsAth+pgif/AYh+2bEhoXYZpwR1Ah7V5aYj2DQASe5ko5CyDt/B6mpBmnd\nHAx+dMBYTko06zlpxvTm6Qa5lBQHxRS5drXUzpJXZLh7MoaCUhIWHiyCHB3cHFwA\no4NdLJaAk69BAn3dF8as/WpRhMIOvSfc78o8ROyiqNYrCEWhIt24Zgbp4pz2ZgS9\nnMJzMhsWwp912S5nbGJ2rEiGJa2wCZEwDCtQiR/1QpGpgMN/Ft6/OD8S2GPp6Gf0\nONdw0jLwAAL6595VEAXBLvCVnOgauOoh3dRUtWT5RxqMoWb3haBE0B6l3BOTLGRH\ndMjFS/z9+HVMRmoYUg+yxt17eF/jPDGhF4s7Tto=\n-----END CERTIFICATE-----\n", "key": "-----BEGIN RSA PRIVATE KEY-----\nMIIJKgIBAAKCAgEAwUdLigUqm4ihe1tN3HWP8iefRCv5BKBZNP7e3i5WDh8Q0XLH\nDYOLEEu7RwfdGYll/Lr7xmJrpPINKZIJ/7Klg4GAcI34fNUCXlSUSZjQ/3fq/YWW\nbit1RbjoNjOWNvzzNBI4m53WCLR4LnTorcQx9NPnaolf3xwXMua9LHbWj3Jc4rnb\nVV7PKLSTyEY07WjdLhqnFRIYnNPjk1e9VRhzq9c8D360FTuEyICa9oLP6KGJxGPX\nfxvk45M4ztEmgdOmmunjRmxRmDXUEj5ZmGJHm75dwjYZsktDeKJVO64+iuLzS9cn\nq5o4ERxpHkLnDAbIZ35YCLiJs8EGV4xygo7xVfYkReLooLHbCCVv9qsfdizi2t/4\nLhxxFpSUGnPaior+Lxic2miu+arZ+2E5EWXPz/QoP65JCfCWDxd7EbDLQ8bcEo3L\nwM6sgmeac184H3bPlfMG/RNn+qI71stdFpgCuM7hD5h+PChEBrhUnCn3fzYTBFHL\nUHwzedVBLKgk9LxecPKIa4vIfShUsD1XhktoaliiKqOq4wgG4z0iJ2wTsIpfJO4R\nTLsGF4pf+cbtpugVmCuMLnH4c4jb2br4D4FwdraGOhBgjoE5zoRdTDh4JOL46M28\nkaO8v4uSB2tVgOkY9aLaHt1urhu9gxNikwzLIvJIS8oRANhiu6Dpbi6c7BkCAwEA\nAQKCAgAK4xzXkVcvM+zxtYjPIW2qdK4pYUc5ih1ASZeNASPuj3QrcilRgu6FTFjI\nJWKgVr+gkvlN+VsKK2rPo8btqPbfD59IJ80engKQRPgl5aZnL450ea8JtWe5ldL5\nw6HjZs/9x8V5yZQmASFwb3wwVpgATCPx2btlZlANuVbgX3DSW30tNc4rssedSb+1\n1TOTyHeGCnENNCvgAKUF9pgnCBWy8TC0SD01YXkHMMxZ1LBIsLnBjuSzh6MW6SL2\n7mvafg05VnGrlLwrZGPV4755AZc6SAjAwPqlUw1ByTYfIp9qBCLzi5Lz3P/pwdWP\ng3H/mM2227RyiE+UaSv16EbGgS4Joh7Y/do2WmAmBS7wFQN/rnywpXBmv5b54xGt\n5IfuqP6Y8z6oBwJ4sP3DM50ReOHL5hzSzl4yJA8kZ+Xk0s7hXEmD2YvlopoHcPPH\n1aiON07Cjf7dLjflX4REg9kbY/+wwxupmxp28LtpTka8tH/a86cy+/ro2PqLLHEy\nDapTFktNdC3J8tOKfO+ceFS1V2GcvF4aqyAb0UxzCI+xnSJC4t+nVO9rNE4Arff2\nrk02yhlSjnnJXIxEc5+jWyE2NuzaUTOLT3+b747a1iaBTtCvBs2zFMDiBK88ZaP9\nuTcQF3Cl45F7JhI3ohCEJpnXe5W1CWcS+UlG1UcrNhG4yAtTYQKCAQEA47sUq80Q\nvcPhdAJord/ZvKvZ9j2ufZt4IW2npCLXhsDDrtbdxGJ0yojAeN2zc0E8KWmPGr5N\nolLrdmY0+nXKayKCx1Z4VYwA1g9KMola+MLgtYCRaTyDLLfaEFH8OWjPdglyOhx9\nTshGB9Kk9ag0ty9rLvsvPP1vlBkIiMYG1SexRe7SJopgtb/PqTv+ftPjmDcll59+\nli/PgkmhEwFQbjJYBTPh3pEk6i8JPV4NvUjPEokApqDfHvsuFIQazmM6mZqv97zT\neHl7R/Ba/KYwwyuGvPcRWCpNqEki5Nj4pZqez9gLS5HXBjsxDtjTmcq4Izr7hxYh\nwZBTlHfWipq1IQKCAQEA2UVgN9SQSRL0SkLYUrHWL/ULzwkdiiLqXz5Fg4bmqJTD\n0ICYiauJDFr1ueKK7jrtJpq9Sb+ytaezwy3UKPUmM12Hn28OK6N7D1kgAftTY748\nKVwSkQkBASkJnMGzbS8tqIa1JQcUv0N1VPWIRM0fe64UVZcl3/aIz+6pgI8rBgmG\nySC2Gv/nNWuHB9nqxGjCxaou0iN6DWz9NX6mEYPkyTDecrLzWEtx9vEjskh4BibC\nk2kclzGLHIZA0jtPk6RSwmgt/jCIqypzDFhQkjM0if1//dwdex4BzF+jk7hq3FVK\nF8hSch69UdjiVhb08jrgU1keEJC2J16XJx7ho/vf+QKCAQEA2uXR2c30gKPPi2Zh\nXnikX9HEYggqOHx/fYpcsM3444NY1S+d3Y5R+vh2Iffg7TFQdfIMZ0VzdEtGH/ng\nhLtChbaXhKpeYsR/bUMnK+qRsjoIKw7ZNxG+ePaEOlnEGx7fA/PSg38tB9REVHKq\ns6MqgMaovXEel/54CSAC7w2CNdPAd5s3D2rFyQ4FuwN6/kgIJiEkFlluIUj1qqPB\nm3YQ7Muhhea3/ur/U8dAl2dXx5E70fI02PYbYjlIstnEtov1IP7qlDRCBTSoRp8H\npoDMsF2fVvYaCugm21LpopXYc75rVxewy8P0rv2oiz/qLZMJanvgmJmyes+TN0v1\nwicxYQKCAQEAwLEZm4pGiLhpte7Awd58sj2mwnA4JYm/yMaGiha7PKnrXGCgaACc\n5DtLkRdM1emm58T2uZBtI1RfRmu/rK92uLlrjcfFjohHnJ4TmA5Yr65I1vmUcdLF\ndEz7/G2B/wSHft4xi2xTRynBxp6Xw0UUdfE2kY4ihhW00oy9AYp3wrvyMTB0sU9h\ns+vwbtMithyfGQVMzwuttHiDvMzc5c3Ou2w+7QSucxE+Op80ZGqHIBWCHZBsTd8f\nnCWpGYqGkc97onUYzUYn/33/WOR+WJtXLkaIHBH6h3VebO4pOdql01FFDACatiFz\nEELEwUtM5AO5bdIsqs5c7l+iuIxX28KHqQKCAQEAqDyoaV6yEFoRrttPuG7UZHPW\no/XdUerB5jSNodlVNkJxZleEt88fm5kO9XPRKvyzBXVGwQPqD1/YrIGzuWmaCknm\nOPJhWvPnOsCS0dyPC9uRBYYNz1X0hBEdZKVorhiSiP08r3SSOnpJPTbLMDI4XPEq\nHGRGQw9Liw8fwLGZY4RFWi0mKiyJDIQTsrxmkavnqRWblV0L1y1ayYZOmyOgDABX\nzsGx6YBUoZNeKIZsbDPy6dr/k8NJvn1PRSGzE9oNffpSf51VwJaMRMvuHrA1elk0\nws6MBfBh5lJ6H70n9Dq5w3S7iSWXAUwqhxSOU+vF18cFYcmu8EfK5SIVB4idZQ==\n-----END RSA PRIVATE KEY-----\n"} 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: cephadm 2026-03-21T06:48:10.169504+0000 mgr.x (mgr.14152) 1 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTING 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: cephadm 2026-03-21T06:48:10.169504+0000 mgr.x (mgr.14152) 1 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTING 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: cephadm 2026-03-21T06:48:10.270812+0000 mgr.x (mgr.14152) 2 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: cephadm 2026-03-21T06:48:10.270812+0000 mgr.x (mgr.14152) 2 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: cephadm 2026-03-21T06:48:10.379369+0000 mgr.x (mgr.14152) 3 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: cephadm 2026-03-21T06:48:10.379369+0000 mgr.x (mgr.14152) 3 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: cephadm 2026-03-21T06:48:10.379416+0000 mgr.x (mgr.14152) 4 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTED 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: cephadm 2026-03-21T06:48:10.379416+0000 mgr.x (mgr.14152) 4 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTED 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: cephadm 2026-03-21T06:48:10.379868+0000 mgr.x (mgr.14152) 5 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Client ('192.168.123.102', 60300) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: cephadm 2026-03-21T06:48:10.379868+0000 mgr.x (mgr.14152) 5 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Client ('192.168.123.102', 60300) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: audit 2026-03-21T06:48:10.539699+0000 mgr.x (mgr.14152) 6 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: audit 2026-03-21T06:48:10.539699+0000 mgr.x (mgr.14152) 6 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: audit 2026-03-21T06:48:10.543919+0000 mgr.x (mgr.14152) 7 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:11.856 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:11 vm02 bash[17657]: audit 2026-03-21T06:48:10.543919+0000 mgr.x (mgr.14152) 7 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:12.141 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout SSL certificate updated 2026-03-21T06:48:12.456 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout SSL certificate key updated 2026-03-21T06:48:12.456 INFO:teuthology.orchestra.run.vm02.stdout:Creating initial admin user... 2026-03-21T06:48:12.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:12 vm02 bash[17657]: audit 2026-03-21T06:48:10.854917+0000 mgr.x (mgr.14152) 8 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "orch certmgr generate-certificates", "module_name": "dashboard", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:12.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:12 vm02 bash[17657]: audit 2026-03-21T06:48:10.854917+0000 mgr.x (mgr.14152) 8 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "orch certmgr generate-certificates", "module_name": "dashboard", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:12.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:12 vm02 bash[17657]: cluster 2026-03-21T06:48:11.593890+0000 mon.a (mon.0) 89 : cluster [DBG] mgrmap e12: x(active, since 2s) 2026-03-21T06:48:12.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:12 vm02 bash[17657]: cluster 2026-03-21T06:48:11.593890+0000 mon.a (mon.0) 89 : cluster [DBG] mgrmap e12: x(active, since 2s) 2026-03-21T06:48:12.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:12 vm02 bash[17657]: audit 2026-03-21T06:48:12.093702+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:12.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:12 vm02 bash[17657]: audit 2026-03-21T06:48:12.093702+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:12.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:12 vm02 bash[17657]: audit 2026-03-21T06:48:12.415641+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:12.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:12 vm02 bash[17657]: audit 2026-03-21T06:48:12.415641+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:12.960 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout {"username": "admin", "password": "$2b$12$fu0j1G/6pJCr1auft9k3Feu/RWO.xbunwE9Bo7VaES3rcyoGvA34C", "roles": ["administrator"], "name": null, "email": null, "lastUpdate": 1774075692, "enabled": true, "pwdExpirationDate": null, "pwdUpdateRequired": true} 2026-03-21T06:48:12.960 INFO:teuthology.orchestra.run.vm02.stdout:Fetching dashboard port number... 2026-03-21T06:48:13.243 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stdout 8443 2026-03-21T06:48:13.243 INFO:teuthology.orchestra.run.vm02.stdout:firewalld does not appear to be present 2026-03-21T06:48:13.243 INFO:teuthology.orchestra.run.vm02.stdout:Not possible to open ports <[8443]>. firewalld.service is not available 2026-03-21T06:48:13.244 INFO:teuthology.orchestra.run.vm02.stdout:Ceph Dashboard is now available at: 2026-03-21T06:48:13.244 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:13.245 INFO:teuthology.orchestra.run.vm02.stdout: URL: https://vm02.local:8443/ 2026-03-21T06:48:13.245 INFO:teuthology.orchestra.run.vm02.stdout: User: admin 2026-03-21T06:48:13.245 INFO:teuthology.orchestra.run.vm02.stdout: Password: jpjt9pixdf 2026-03-21T06:48:13.245 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:13.245 INFO:teuthology.orchestra.run.vm02.stdout:Saving cluster configuration to /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config directory 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout:/usr/bin/ceph: stderr set mgr/dashboard/cluster/status 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout:You can access the Ceph CLI as following in case of multi-cluster or non-default config: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: sudo /home/ubuntu/cephtest/cephadm shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout:Or, if you are only running a single cluster on this host: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: sudo /home/ubuntu/cephtest/cephadm shell 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout:Please consider enabling telemetry to help improve Ceph: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: ceph telemetry on 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout:For more information see: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: https://docs.ceph.com/en/latest/mgr/telemetry/ 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:13.597 INFO:teuthology.orchestra.run.vm02.stdout:Bootstrap complete. 2026-03-21T06:48:13.603 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout static 2026-03-21T06:48:13.605 INFO:teuthology.orchestra.run.vm02.stdout:Non-zero exit code 3 from systemctl is-active logrotate 2026-03-21T06:48:13.605 INFO:teuthology.orchestra.run.vm02.stdout:systemctl: stdout inactive 2026-03-21T06:48:13.605 INFO:teuthology.orchestra.run.vm02.stdout:Enabling the logrotate.timer service to perform daily log rotation. 2026-03-21T06:48:13.815 INFO:tasks.cephadm:Fetching config... 2026-03-21T06:48:13.815 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:48:13.815 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.conf of=/dev/stdout 2026-03-21T06:48:13.818 INFO:tasks.cephadm:Fetching client.admin keyring... 2026-03-21T06:48:13.818 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:48:13.818 DEBUG:teuthology.orchestra.run.vm02:> dd if=/etc/ceph/ceph.client.admin.keyring of=/dev/stdout 2026-03-21T06:48:13.862 INFO:tasks.cephadm:Fetching mon keyring... 2026-03-21T06:48:13.862 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:48:13.862 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/keyring of=/dev/stdout 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:13 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:13 vm02 bash[17657]: audit 2026-03-21T06:48:12.089751+0000 mgr.x (mgr.14152) 9 : audit [DBG] from='client.14166 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:13 vm02 bash[17657]: audit 2026-03-21T06:48:12.089751+0000 mgr.x (mgr.14152) 9 : audit [DBG] from='client.14166 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:13 vm02 bash[17657]: audit 2026-03-21T06:48:12.411813+0000 mgr.x (mgr.14152) 10 : audit [DBG] from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:13 vm02 bash[17657]: audit 2026-03-21T06:48:12.411813+0000 mgr.x (mgr.14152) 10 : audit [DBG] from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:13 vm02 bash[17657]: audit 2026-03-21T06:48:12.916726+0000 mon.a (mon.0) 92 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:13 vm02 bash[17657]: audit 2026-03-21T06:48:12.916726+0000 mon.a (mon.0) 92 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:13 vm02 bash[17657]: audit 2026-03-21T06:48:13.203820+0000 mon.a (mon.0) 93 : audit [DBG] from='client.? 192.168.123.102:0/2434461531' entity='client.admin' cmd={"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"} : dispatch 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:13 vm02 bash[17657]: audit 2026-03-21T06:48:13.203820+0000 mon.a (mon.0) 93 : audit [DBG] from='client.? 192.168.123.102:0/2434461531' entity='client.admin' cmd={"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"} : dispatch 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:13 vm02 bash[17657]: audit 2026-03-21T06:48:13.555919+0000 mon.a (mon.0) 94 : audit [INF] from='client.? 192.168.123.102:0/721893450' entity='client.admin' 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:13 vm02 bash[17657]: audit 2026-03-21T06:48:13.555919+0000 mon.a (mon.0) 94 : audit [INF] from='client.? 192.168.123.102:0/721893450' entity='client.admin' 2026-03-21T06:48:13.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:13 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:13.902 INFO:tasks.cephadm:Fetching pub ssh key... 2026-03-21T06:48:13.902 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:48:13.902 DEBUG:teuthology.orchestra.run.vm02:> dd if=/home/ubuntu/cephtest/ceph.pub of=/dev/stdout 2026-03-21T06:48:13.950 INFO:tasks.cephadm:Installing pub ssh key for root users... 2026-03-21T06:48:13.950 DEBUG:teuthology.orchestra.run.vm02:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDqVQU8deI5GkrkKLViWR7HEg9MGwqno6Oh6IzEbn/fy ceph-b16ecafc-24f1-11f1-8ede-8330751617ee' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-21T06:48:14.006 INFO:teuthology.orchestra.run.vm02.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDqVQU8deI5GkrkKLViWR7HEg9MGwqno6Oh6IzEbn/fy ceph-b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:14.012 DEBUG:teuthology.orchestra.run.vm04:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDqVQU8deI5GkrkKLViWR7HEg9MGwqno6Oh6IzEbn/fy ceph-b16ecafc-24f1-11f1-8ede-8330751617ee' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-21T06:48:14.023 INFO:teuthology.orchestra.run.vm04.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDqVQU8deI5GkrkKLViWR7HEg9MGwqno6Oh6IzEbn/fy ceph-b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:14.028 DEBUG:teuthology.orchestra.run.vm07:> sudo install -d -m 0700 /root/.ssh && echo 'ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDqVQU8deI5GkrkKLViWR7HEg9MGwqno6Oh6IzEbn/fy ceph-b16ecafc-24f1-11f1-8ede-8330751617ee' | sudo tee -a /root/.ssh/authorized_keys && sudo chmod 0600 /root/.ssh/authorized_keys 2026-03-21T06:48:14.039 INFO:teuthology.orchestra.run.vm07.stdout:ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDqVQU8deI5GkrkKLViWR7HEg9MGwqno6Oh6IzEbn/fy ceph-b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:14.043 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph config set mgr mgr/cephadm/allow_ptrace true 2026-03-21T06:48:14.286 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:48:14.670 INFO:tasks.cephadm:Distributing conf and client.admin keyring to all hosts + 0755 2026-03-21T06:48:14.670 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch client-keyring set client.admin '*' --mode 0755 2026-03-21T06:48:14.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:14 vm02 bash[17657]: audit 2026-03-21T06:48:12.711292+0000 mgr.x (mgr.14152) 11 : audit [DBG] from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:14.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:14 vm02 bash[17657]: audit 2026-03-21T06:48:12.711292+0000 mgr.x (mgr.14152) 11 : audit [DBG] from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:14.916 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:48:15.345 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm04 2026-03-21T06:48:15.345 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:48:15.345 DEBUG:teuthology.orchestra.run.vm04:> dd of=/etc/ceph/ceph.conf 2026-03-21T06:48:15.349 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:48:15.349 DEBUG:teuthology.orchestra.run.vm04:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:15.395 INFO:tasks.cephadm:Adding host vm04 to orchestrator... 2026-03-21T06:48:15.395 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch host add vm04 2026-03-21T06:48:15.646 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:48:15.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:15 vm02 bash[17657]: audit 2026-03-21T06:48:14.608592+0000 mon.a (mon.0) 95 : audit [INF] from='client.? 192.168.123.102:0/1702136911' entity='client.admin' 2026-03-21T06:48:15.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:15 vm02 bash[17657]: audit 2026-03-21T06:48:14.608592+0000 mon.a (mon.0) 95 : audit [INF] from='client.? 192.168.123.102:0/1702136911' entity='client.admin' 2026-03-21T06:48:15.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:15 vm02 bash[17657]: audit 2026-03-21T06:48:14.697746+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:15.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:15 vm02 bash[17657]: audit 2026-03-21T06:48:14.697746+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:15.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:15 vm02 bash[17657]: audit 2026-03-21T06:48:15.266639+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:15.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:15 vm02 bash[17657]: audit 2026-03-21T06:48:15.266639+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:15.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:15 vm02 bash[17657]: audit 2026-03-21T06:48:15.271954+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:15.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:15 vm02 bash[17657]: audit 2026-03-21T06:48:15.271954+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.616 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:15.268547+0000 mgr.x (mgr.14152) 12 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:16.616 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:15.268547+0000 mgr.x (mgr.14152) 12 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: cluster 2026-03-21T06:48:16.275985+0000 mon.a (mon.0) 99 : cluster [DBG] mgrmap e13: x(active, since 6s) 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: cluster 2026-03-21T06:48:16.275985+0000 mon.a (mon.0) 99 : cluster [DBG] mgrmap e13: x(active, since 6s) 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.288184+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.288184+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.291087+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.291087+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.291911+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.291911+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.292731+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.292731+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.293263+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.293263+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.416070+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.416070+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.419237+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.419237+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.421755+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.421755+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.426463+0000 mon.a (mon.0) 108 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.426463+0000 mon.a (mon.0) 108 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.427256+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.427256+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.427695+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.427695+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.430299+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:16 vm02 bash[17657]: audit 2026-03-21T06:48:16.430299+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:17.832 INFO:teuthology.orchestra.run.vm02.stdout:Added host 'vm04' with addr '192.168.123.104' 2026-03-21T06:48:17.889 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch host ls --format=json 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: audit 2026-03-21T06:48:15.957062+0000 mgr.x (mgr.14152) 13 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: audit 2026-03-21T06:48:15.957062+0000 mgr.x (mgr.14152) 13 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: cephadm 2026-03-21T06:48:16.294027+0000 mgr.x (mgr.14152) 14 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: cephadm 2026-03-21T06:48:16.294027+0000 mgr.x (mgr.14152) 14 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: cephadm 2026-03-21T06:48:16.325888+0000 mgr.x (mgr.14152) 15 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: cephadm 2026-03-21T06:48:16.325888+0000 mgr.x (mgr.14152) 15 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: cephadm 2026-03-21T06:48:16.354553+0000 mgr.x (mgr.14152) 16 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: cephadm 2026-03-21T06:48:16.354553+0000 mgr.x (mgr.14152) 16 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: cephadm 2026-03-21T06:48:16.381714+0000 mgr.x (mgr.14152) 17 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: cephadm 2026-03-21T06:48:16.381714+0000 mgr.x (mgr.14152) 17 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: cephadm 2026-03-21T06:48:16.503567+0000 mgr.x (mgr.14152) 18 : cephadm [INF] Deploying cephadm binary to vm04 2026-03-21T06:48:17.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:17 vm02 bash[17657]: cephadm 2026-03-21T06:48:16.503567+0000 mgr.x (mgr.14152) 18 : cephadm [INF] Deploying cephadm binary to vm04 2026-03-21T06:48:18.126 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:48:18.439 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:18.439 INFO:teuthology.orchestra.run.vm02.stdout:[{"addr": "192.168.123.102", "hostname": "vm02", "labels": [], "status": ""}, {"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}] 2026-03-21T06:48:18.492 INFO:tasks.cephadm:Writing (initial) conf and keyring to vm07 2026-03-21T06:48:18.492 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:48:18.492 DEBUG:teuthology.orchestra.run.vm07:> dd of=/etc/ceph/ceph.conf 2026-03-21T06:48:18.495 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:48:18.495 DEBUG:teuthology.orchestra.run.vm07:> dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:18.541 INFO:tasks.cephadm:Adding host vm07 to orchestrator... 2026-03-21T06:48:18.541 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch host add vm07 2026-03-21T06:48:18.788 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: audit 2026-03-21T06:48:17.825992+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: audit 2026-03-21T06:48:17.825992+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: cephadm 2026-03-21T06:48:17.826396+0000 mgr.x (mgr.14152) 19 : cephadm [INF] Added host vm04 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: cephadm 2026-03-21T06:48:17.826396+0000 mgr.x (mgr.14152) 19 : cephadm [INF] Added host vm04 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: audit 2026-03-21T06:48:17.826677+0000 mon.a (mon.0) 113 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: audit 2026-03-21T06:48:17.826677+0000 mon.a (mon.0) 113 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: audit 2026-03-21T06:48:18.113003+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: audit 2026-03-21T06:48:18.113003+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: audit 2026-03-21T06:48:18.368400+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: audit 2026-03-21T06:48:18.368400+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: audit 2026-03-21T06:48:18.433860+0000 mgr.x (mgr.14152) 20 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:18 vm02 bash[17657]: audit 2026-03-21T06:48:18.433860+0000 mgr.x (mgr.14152) 20 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:20.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:19 vm02 bash[17657]: audit 2026-03-21T06:48:18.928895+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:20.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:19 vm02 bash[17657]: audit 2026-03-21T06:48:18.928895+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:20.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:19 vm02 bash[17657]: audit 2026-03-21T06:48:19.104814+0000 mgr.x (mgr.14152) 21 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm07", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:20.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:19 vm02 bash[17657]: audit 2026-03-21T06:48:19.104814+0000 mgr.x (mgr.14152) 21 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm07", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:21.149 INFO:teuthology.orchestra.run.vm02.stdout:Added host 'vm07' with addr '192.168.123.107' 2026-03-21T06:48:21.225 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch host ls --format=json 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: cephadm 2026-03-21T06:48:19.708613+0000 mgr.x (mgr.14152) 22 : cephadm [INF] Deploying cephadm binary to vm07 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: cephadm 2026-03-21T06:48:19.708613+0000 mgr.x (mgr.14152) 22 : cephadm [INF] Deploying cephadm binary to vm07 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.029193+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.029193+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.031967+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.031967+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.035357+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.035357+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.038086+0000 mon.a (mon.0) 120 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.038086+0000 mon.a (mon.0) 120 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.038646+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.038646+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.039370+0000 mon.a (mon.0) 122 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.039370+0000 mon.a (mon.0) 122 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.039897+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.039897+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: cephadm 2026-03-21T06:48:20.040563+0000 mgr.x (mgr.14152) 23 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: cephadm 2026-03-21T06:48:20.040563+0000 mgr.x (mgr.14152) 23 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: cephadm 2026-03-21T06:48:20.080611+0000 mgr.x (mgr.14152) 24 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: cephadm 2026-03-21T06:48:20.080611+0000 mgr.x (mgr.14152) 24 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: cephadm 2026-03-21T06:48:20.116690+0000 mgr.x (mgr.14152) 25 : cephadm [INF] Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: cephadm 2026-03-21T06:48:20.116690+0000 mgr.x (mgr.14152) 25 : cephadm [INF] Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: cephadm 2026-03-21T06:48:20.150701+0000 mgr.x (mgr.14152) 26 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: cephadm 2026-03-21T06:48:20.150701+0000 mgr.x (mgr.14152) 26 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.615400+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.615400+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.619283+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.619283+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.622655+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:21 vm02 bash[17657]: audit 2026-03-21T06:48:20.622655+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:21.460 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:48:21.790 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:21.790 INFO:teuthology.orchestra.run.vm02.stdout:[{"addr": "192.168.123.102", "hostname": "vm02", "labels": [], "status": ""}, {"addr": "192.168.123.104", "hostname": "vm04", "labels": [], "status": ""}, {"addr": "192.168.123.107", "hostname": "vm07", "labels": [], "status": ""}] 2026-03-21T06:48:21.870 INFO:tasks.cephadm:Setting crush tunables to default 2026-03-21T06:48:21.870 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd crush tunables default 2026-03-21T06:48:22.102 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:48:22.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:22 vm02 bash[17657]: audit 2026-03-21T06:48:21.143323+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:22.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:22 vm02 bash[17657]: audit 2026-03-21T06:48:21.143323+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:22.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:22 vm02 bash[17657]: cephadm 2026-03-21T06:48:21.143828+0000 mgr.x (mgr.14152) 27 : cephadm [INF] Added host vm07 2026-03-21T06:48:22.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:22 vm02 bash[17657]: cephadm 2026-03-21T06:48:21.143828+0000 mgr.x (mgr.14152) 27 : cephadm [INF] Added host vm07 2026-03-21T06:48:22.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:22 vm02 bash[17657]: audit 2026-03-21T06:48:21.144202+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:22.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:22 vm02 bash[17657]: audit 2026-03-21T06:48:21.144202+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:22.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:22 vm02 bash[17657]: audit 2026-03-21T06:48:21.440241+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:22.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:22 vm02 bash[17657]: audit 2026-03-21T06:48:21.440241+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:22.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:22 vm02 bash[17657]: audit 2026-03-21T06:48:21.706038+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:22.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:22 vm02 bash[17657]: audit 2026-03-21T06:48:21.706038+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:23.290 INFO:teuthology.orchestra.run.vm02.stderr:adjusted tunables profile to default 2026-03-21T06:48:23.348 INFO:tasks.cephadm:Adding mon.a on vm02 2026-03-21T06:48:23.348 INFO:tasks.cephadm:Adding mon.b on vm04 2026-03-21T06:48:23.348 INFO:tasks.cephadm:Adding mon.c on vm07 2026-03-21T06:48:23.348 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch apply mon '3;vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c' 2026-03-21T06:48:23.582 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:21.784850+0000 mgr.x (mgr.14152) 28 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:21.784850+0000 mgr.x (mgr.14152) 28 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:22.280754+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:22.280754+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:22.411837+0000 mon.a (mon.0) 132 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd={"prefix": "osd crush tunables", "profile": "default"} : dispatch 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:22.411837+0000 mon.a (mon.0) 132 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd={"prefix": "osd crush tunables", "profile": "default"} : dispatch 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.229669+0000 mon.a (mon.0) 133 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.229669+0000 mon.a (mon.0) 133 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.232858+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.232858+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.236368+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.236368+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.239045+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.239045+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.239534+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.239534+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.240204+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.240204+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.240646+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:23.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:23 vm02 bash[17657]: audit 2026-03-21T06:48:23.240646+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:23.901 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled mon update... 2026-03-21T06:48:23.980 DEBUG:teuthology.orchestra.run.vm04:mon.b> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.b.service 2026-03-21T06:48:23.981 DEBUG:teuthology.orchestra.run.vm07:mon.c> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.c.service 2026-03-21T06:48:23.982 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-21T06:48:23.982 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph mon dump -f json 2026-03-21T06:48:24.260 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:24.637 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:48:24.637 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":1,"fsid":"b16ecafc-24f1-11f1-8ede-8330751617ee","modified":"2026-03-21T06:47:46.690400Z","created":"2026-03-21T06:47:46.690400Z","min_mon_release":20,"min_mon_release_name":"tentacle","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid","tentacle","nvmeof_beacon_diff"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0]} 2026-03-21T06:48:24.637 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 1 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.241301+0000 mgr.x (mgr.14152) 29 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.241301+0000 mgr.x (mgr.14152) 29 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.271217+0000 mgr.x (mgr.14152) 30 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.271217+0000 mgr.x (mgr.14152) 30 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.285120+0000 mon.a (mon.0) 140 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.285120+0000 mon.a (mon.0) 140 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: cluster 2026-03-21T06:48:23.289349+0000 mon.a (mon.0) 141 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: cluster 2026-03-21T06:48:23.289349+0000 mon.a (mon.0) 141 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.303461+0000 mgr.x (mgr.14152) 31 : cephadm [INF] Updating vm07:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.303461+0000 mgr.x (mgr.14152) 31 : cephadm [INF] Updating vm07:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.335772+0000 mgr.x (mgr.14152) 32 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.335772+0000 mgr.x (mgr.14152) 32 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.369557+0000 mon.a (mon.0) 142 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.369557+0000 mon.a (mon.0) 142 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.372382+0000 mon.a (mon.0) 143 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.372382+0000 mon.a (mon.0) 143 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.374780+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.374780+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.895775+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.895775+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.896325+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.896325+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.897294+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.897294+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.897802+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.897802+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.901129+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.901129+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.902098+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.902098+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.902507+0000 mon.a (mon.0) 151 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:24.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:24 vm02 bash[17657]: audit 2026-03-21T06:48:23.902507+0000 mon.a (mon.0) 151 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.089 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:25.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:25.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 systemd[1]: Started Ceph mon.c for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 0 set uid:gid to 167:167 (ceph:ceph) 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 0 ceph version 20.2.0-712-g70f8415b (70f8415b300f041766fa27faf7d5472699e32388) tentacle (stable - Debug), process ceph-mon, pid 7 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 0 pidfile_write: ignore empty --pid-file 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 0 load: jerasure load: lrc 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: RocksDB version: 7.9.2 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Git sha 0 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Compile date 2026-03-13 17:47:44 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: DB SUMMARY 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: DB Session ID: D2QYLJ00KI020PRJF0UA 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: CURRENT file: CURRENT 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: IDENTITY file: IDENTITY 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: MANIFEST file: MANIFEST-000005 size: 59 Bytes 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: SST files in /var/lib/ceph/mon/ceph-c/store.db dir, Total Num: 0, files: 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Write Ahead Log file in /var/lib/ceph/mon/ceph-c/store.db: 000004.log size: 511 ; 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.error_if_exists: 0 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.create_if_missing: 0 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.paranoid_checks: 1 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.flush_verify_memtable_count: 1 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.track_and_verify_wals_in_manifest: 0 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.verify_sst_unique_id_in_manifest: 1 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.env: 0x55913508e100 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.fs: PosixFileSystem 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.info_log: 0x559168243b00 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_file_opening_threads: 16 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.statistics: (nil) 2026-03-21T06:48:25.437 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.use_fsync: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_log_file_size: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_manifest_file_size: 1073741824 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.log_file_time_to_roll: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.keep_log_file_num: 1000 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.recycle_log_file_num: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.allow_fallocate: 1 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.allow_mmap_reads: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.allow_mmap_writes: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.use_direct_reads: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.use_direct_io_for_flush_and_compaction: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.create_missing_column_families: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.db_log_dir: 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.wal_dir: 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.table_cache_numshardbits: 6 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.WAL_ttl_seconds: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.WAL_size_limit_MB: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_write_batch_group_size_bytes: 1048576 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.manifest_preallocation_size: 4194304 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.is_fd_close_on_exec: 1 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.advise_random_on_open: 1 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.db_write_buffer_size: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.write_buffer_manager: 0x559168246500 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.access_hint_on_compaction_start: 1 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.random_access_max_buffer_size: 1048576 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.use_adaptive_mutex: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.rate_limiter: (nil) 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.sst_file_manager.rate_bytes_per_sec: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.wal_recovery_mode: 2 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.enable_thread_tracking: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.enable_pipelined_write: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.unordered_write: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.allow_concurrent_memtable_write: 1 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.enable_write_thread_adaptive_yield: 1 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.write_thread_max_yield_usec: 100 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.write_thread_slow_yield_usec: 3 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.row_cache: None 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.wal_filter: None 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.avoid_flush_during_recovery: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.allow_ingest_behind: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.two_write_queues: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.manual_wal_flush: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.wal_compression: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.atomic_flush: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.avoid_unnecessary_blocking_io: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.persist_stats_to_disk: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.write_dbid_to_manifest: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.log_readahead_size: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.file_checksum_gen_factory: Unknown 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.best_efforts_recovery: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_bgerror_resume_count: 2147483647 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bgerror_resume_retry_interval: 1000000 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.allow_data_in_errors: 0 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.db_host_id: __hostname__ 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.enforce_single_del_contracts: true 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_background_jobs: 2 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_background_compactions: -1 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_subcompactions: 1 2026-03-21T06:48:25.438 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.avoid_flush_during_shutdown: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.writable_file_max_buffer_size: 1048576 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.delayed_write_rate : 16777216 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_total_wal_size: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.delete_obsolete_files_period_micros: 21600000000 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.stats_dump_period_sec: 600 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.stats_persist_period_sec: 600 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.stats_history_buffer_size: 1048576 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_open_files: -1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bytes_per_sync: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.wal_bytes_per_sync: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.strict_bytes_per_sync: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_readahead_size: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_background_flushes: -1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Compression algorithms supported: 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: kZSTD supported: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: kXpressCompression supported: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: kBZip2Compression supported: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: kZSTDNotFinalCompression supported: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: kLZ4Compression supported: 1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: kZlibCompression supported: 1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: kLZ4HCCompression supported: 1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: kSnappyCompression supported: 1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Fast CRC32 supported: Supported on x86 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: DMutex implementation: pthread_mutex_t 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: [db/version_set.cc:5527] Recovering from manifest file: /var/lib/ceph/mon/ceph-c/store.db/MANIFEST-000005 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: [db/column_family.cc:630] --------------- Options for column family [default]: 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.comparator: leveldb.BytewiseComparator 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.merge_operator: 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_filter: None 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_filter_factory: None 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.sst_partitioner_factory: None 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.memtable_factory: SkipListFactory 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.table_factory: BlockBasedTable 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: table_factory options: flush_block_policy_factory: FlushBlockBySizePolicyFactory (0x559168242240) 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cache_index_and_filter_blocks: 1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cache_index_and_filter_blocks_with_high_priority: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: pin_l0_filter_and_index_blocks_in_cache: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: pin_top_level_index_and_filter: 1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: index_type: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: data_block_index_type: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: index_shortening: 1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: data_block_hash_table_util_ratio: 0.750000 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: checksum: 4 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: no_block_cache: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: block_cache: 0x559168263b90 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: block_cache_name: BinnedLRUCache 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: block_cache_options: 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: capacity : 536870912 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: num_shard_bits : 4 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: strict_capacity_limit : 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: high_pri_pool_ratio: 0.000 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: block_cache_compressed: (nil) 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: persistent_cache: (nil) 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: block_size: 4096 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: block_size_deviation: 10 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: block_restart_interval: 16 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: index_block_restart_interval: 1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: metadata_block_size: 4096 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: partition_filters: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: use_delta_encoding: 1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: filter_policy: bloomfilter 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: whole_key_filtering: 1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: verify_compression: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: read_amp_bytes_per_bit: 0 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: format_version: 5 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: enable_index_compression: 1 2026-03-21T06:48:25.439 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: block_align: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: max_auto_readahead_size: 262144 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: prepopulate_block_cache: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: initial_auto_readahead_size: 8192 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: num_file_reads_for_auto_readahead: 2 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.write_buffer_size: 33554432 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_write_buffer_number: 2 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compression: NoCompression 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bottommost_compression: Disabled 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.prefix_extractor: nullptr 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.memtable_insert_with_hint_prefix_extractor: nullptr 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.num_levels: 7 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.min_write_buffer_number_to_merge: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_write_buffer_number_to_maintain: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_write_buffer_size_to_maintain: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bottommost_compression_opts.window_bits: -14 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bottommost_compression_opts.level: 32767 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bottommost_compression_opts.strategy: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_bytes: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bottommost_compression_opts.zstd_max_train_bytes: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bottommost_compression_opts.parallel_threads: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bottommost_compression_opts.enabled: false 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bottommost_compression_opts.max_dict_buffer_bytes: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bottommost_compression_opts.use_zstd_dict_trainer: true 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compression_opts.window_bits: -14 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compression_opts.level: 32767 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compression_opts.strategy: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compression_opts.max_dict_bytes: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compression_opts.zstd_max_train_bytes: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compression_opts.use_zstd_dict_trainer: true 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compression_opts.parallel_threads: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compression_opts.enabled: false 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compression_opts.max_dict_buffer_bytes: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.level0_file_num_compaction_trigger: 4 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.level0_slowdown_writes_trigger: 20 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.level0_stop_writes_trigger: 36 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.target_file_size_base: 67108864 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.target_file_size_multiplier: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_bytes_for_level_base: 268435456 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.level_compaction_dynamic_level_bytes: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_bytes_for_level_multiplier: 10.000000 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[0]: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[1]: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[2]: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[3]: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[4]: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[5]: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_bytes_for_level_multiplier_addtl[6]: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_sequential_skip_in_iterations: 8 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_compaction_bytes: 1677721600 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.ignore_max_compaction_bytes_for_input: true 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.arena_block_size: 1048576 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.soft_pending_compaction_bytes_limit: 68719476736 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.hard_pending_compaction_bytes_limit: 274877906944 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.disable_auto_compactions: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_style: kCompactionStyleLevel 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_pri: kMinOverlappingRatio 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_options_universal.size_ratio: 1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_options_universal.min_merge_width: 2 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_options_universal.max_merge_width: 4294967295 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_options_universal.max_size_amplification_percent: 200 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_options_universal.compression_size_percent: -1 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_options_universal.stop_style: kCompactionStopStyleTotalSize 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_options_fifo.max_table_files_size: 1073741824 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.compaction_options_fifo.allow_compaction: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.table_properties_collectors: CompactOnDeletionCollector (Sliding window size = 32768 Deletion trigger = 16384 Deletion ratio = 0); 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.inplace_update_support: 0 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.inplace_update_num_locks: 10000 2026-03-21T06:48:25.440 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.memtable_prefix_bloom_size_ratio: 0.000000 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.memtable_whole_key_filtering: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.memtable_huge_page_size: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.bloom_locality: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.max_successive_merges: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.optimize_filters_for_hits: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.paranoid_file_checks: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.force_consistency_checks: 1 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.report_bg_io_stats: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.ttl: 2592000 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.periodic_compaction_seconds: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.preclude_last_level_data_seconds: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.preserve_internal_time_seconds: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.enable_blob_files: false 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.min_blob_size: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.blob_file_size: 268435456 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.blob_compression_type: NoCompression 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.enable_blob_garbage_collection: false 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.blob_garbage_collection_age_cutoff: 0.250000 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.blob_garbage_collection_force_threshold: 1.000000 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.blob_compaction_readahead_size: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.blob_file_starting_level: 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.361+0000 7f7b718d5d80 4 rocksdb: Options.experimental_mempurge_threshold: 0.000000 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.365+0000 7f7b718d5d80 4 rocksdb: [db/version_set.cc:5566] Recovered from manifest file:/var/lib/ceph/mon/ceph-c/store.db/MANIFEST-000005 succeeded,manifest_file_number is 5, next_file_number is 7, last_sequence is 0, log_number is 0,prev_log_number is 0,max_column_family is 0,min_log_number_to_keep is 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.365+0000 7f7b718d5d80 4 rocksdb: [db/version_set.cc:5581] Column family [default] (ID 0), log number is 0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.365+0000 7f7b718d5d80 4 rocksdb: [db/db_impl/db_impl_open.cc:539] DB ID: cb05310f-3656-4ce9-ba90-9f8fef111774 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.365+0000 7f7b718d5d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774075705368130, "job": 1, "event": "recovery_started", "wal_files": [4]} 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.365+0000 7f7b718d5d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1043] Recovering log #4 mode 2 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.365+0000 7f7b718d5d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774075705368896, "cf_name": "default", "job": 1, "event": "table_file_creation", "file_number": 8, "file_size": 1643, "file_checksum": "", "file_checksum_func_name": "Unknown", "smallest_seqno": 1, "largest_seqno": 5, "table_properties": {"data_size": 523, "index_size": 31, "index_partitions": 0, "top_level_index_size": 0, "index_key_is_user_key": 1, "index_value_is_delta_encoded": 1, "filter_size": 69, "raw_key_size": 115, "raw_average_key_size": 23, "raw_value_size": 401, "raw_average_value_size": 80, "num_data_blocks": 1, "num_entries": 5, "num_filter_entries": 5, "num_deletions": 0, "num_merge_operands": 0, "num_range_deletions": 0, "format_version": 0, "fixed_key_len": 0, "filter_policy": "bloomfilter", "column_family_name": "default", "column_family_id": 0, "comparator": "leveldb.BytewiseComparator", "merge_operator": "", "prefix_extractor_name": "nullptr", "property_collectors": "[CompactOnDeletionCollector]", "compression": "NoCompression", "compression_options": "window_bits=-14; level=32767; strategy=0; max_dict_bytes=0; zstd_max_train_bytes=0; enabled=0; max_dict_buffer_bytes=0; use_zstd_dict_trainer=1; ", "creation_time": 1774075705, "oldest_key_time": 0, "file_creation_time": 0, "slow_compression_estimated_data_size": 0, "fast_compression_estimated_data_size": 0, "db_id": "cb05310f-3656-4ce9-ba90-9f8fef111774", "db_session_id": "D2QYLJ00KI020PRJF0UA", "orig_file_number": 8, "seqno_to_time_mapping": "N/A"}} 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.365+0000 7f7b718d5d80 4 rocksdb: EVENT_LOG_v1 {"time_micros": 1774075705368949, "job": 1, "event": "recovery_finished"} 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.365+0000 7f7b718d5d80 4 rocksdb: [db/version_set.cc:5047] Creating manifest 10 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.369+0000 7f7b718d5d80 4 rocksdb: [file/delete_scheduler.cc:74] Deleted file /var/lib/ceph/mon/ceph-c/store.db/000004.log immediately, rate_bytes_per_sec 0, total_trash_size 0 max_trash_db_ratio 0.250000 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.369+0000 7f7b718d5d80 4 rocksdb: [db/db_impl/db_impl_open.cc:1987] SstFileManager instance 0x559168266700 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.369+0000 7f7b718d5d80 4 rocksdb: DB pointer 0x5591682b0000 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.369+0000 7f7b6765e640 4 rocksdb: [db/db_impl/db_impl.cc:1109] ------- DUMPING STATS ------- 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.369+0000 7f7b6765e640 4 rocksdb: [db/db_impl/db_impl.cc:1111] 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: ** DB Stats ** 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Cumulative writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 GB, 0.00 MB/s 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Cumulative WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Cumulative stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Interval writes: 0 writes, 0 keys, 0 commit groups, 0.0 writes per commit group, ingest: 0.00 MB, 0.00 MB/s 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Interval WAL: 0 writes, 0 syncs, 0.00 writes per sync, written: 0.00 GB, 0.00 MB/s 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Interval stall: 00:00:0.000 H:M:S, 0.0 percent 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: ** Compaction Stats [default] ** 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Level Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: L0 1/0 1.60 KB 0.2 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.1 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Sum 1/0 1.60 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.1 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Int 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 1.0 0.0 2.1 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-21T06:48:25.441 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: ** Compaction Stats [default] ** 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Priority Files Size Score Read(GB) Rn(GB) Rnp1(GB) Write(GB) Wnew(GB) Moved(GB) W-Amp Rd(MB/s) Wr(MB/s) Comp(sec) CompMergeCPU(sec) Comp(cnt) Avg(sec) KeyIn KeyDrop Rblob(GB) Wblob(GB) 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: --------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: User 0/0 0.00 KB 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 2.1 0.00 0.00 1 0.001 0 0 0.0 0.0 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Blob file count: 0, total size: 0.0 GB, garbage size: 0.0 GB, space amp: 0.0 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Uptime(secs): 0.0 total, 0.0 interval 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Flush(GB): cumulative 0.000, interval 0.000 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: AddFile(GB): cumulative 0.000, interval 0.000 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: AddFile(Total Files): cumulative 0, interval 0 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: AddFile(L0 Files): cumulative 0, interval 0 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: AddFile(Keys): cumulative 0, interval 0 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Cumulative compaction: 0.00 GB write, 0.30 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Interval compaction: 0.00 GB write, 0.30 MB/s write, 0.00 GB read, 0.00 MB/s read, 0.0 seconds 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Stalls(count): 0 level0_slowdown, 0 level0_slowdown_with_compaction, 0 level0_numfiles, 0 level0_numfiles_with_compaction, 0 stop for pending_compaction_bytes, 0 slowdown for pending_compaction_bytes, 0 memtable_compaction, 0 memtable_slowdown, interval 0 total count 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Block cache BinnedLRUCache@0x559168263b90#7 capacity: 512.00 MB usage: 0.23 KB table_size: 0 occupancy: 18446744073709551615 collections: 1 last_copies: 0 last_secs: 8e-06 secs_since: 0 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: Block cache entry stats(count,size,portion): FilterBlock(1,0.12 KB,2.23517e-05%) IndexBlock(1,0.11 KB,2.08616e-05%) Misc(1,0.00 KB,0%) 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: ** File Read Latency Histogram By Level [default] ** 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.369+0000 7f7b718d5d80 0 mon.c does not exist in monmap, will attempt to join an existing cluster 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.369+0000 7f7b718d5d80 0 using public_addr v2:192.168.123.107:0/0 -> [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.369+0000 7f7b718d5d80 0 starting mon.c rank -1 at public addrs [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] at bind addrs [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon_data /var/lib/ceph/mon/ceph-c fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.369+0000 7f7b718d5d80 1 mon.c@-1(???) e0 preinit fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 0 mon.c@-1(synchronizing).mds e1 new map 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 0 mon.c@-1(synchronizing).mds e1 print_map 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: e1 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: btime 2026-03-21T06:47:47:923913+0000 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: legacy client fscid: -1 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: No filesystems configured 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 1 mon.c@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 1 mon.c@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 1 mon.c@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 1 mon.c@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 1 mon.c@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 1 mon.c@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 0 mon.c@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 0 mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 0 mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.389+0000 7f7b6a664640 0 mon.c@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:47.924339+0000 mon.a (mon.0) 0 : cluster [INF] mkfs b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:47.924339+0000 mon.a (mon.0) 0 : cluster [INF] mkfs b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:47.919314+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:47.919314+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920244+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920244+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920278+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920278+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920283+0000 mon.a (mon.0) 3 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920283+0000 mon.a (mon.0) 3 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920287+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920287+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920297+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920297+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:25.442 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920302+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920302+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920311+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920311+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920314+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920314+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920633+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920633+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920648+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.920648+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.921256+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:48.921256+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:48.990782+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1389498430' entity='client.admin' 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:48.990782+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1389498430' entity='client.admin' 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:49.586159+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/4046014269' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:49.586159+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/4046014269' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:51.810831+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/788823380' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:51.810831+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/788823380' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:52.480256+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:52.480256+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:52.484714+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: x(active, starting, since 0.00461082s) 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:52.484714+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: x(active, starting, since 0.00461082s) 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.486227+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.486227+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.486600+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.486600+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.486956+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.486956+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.487347+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.487347+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.487688+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.487688+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:52.494572+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:52.494572+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.503203+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.503203+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.507496+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.507496+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.507813+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.507813+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.511606+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.511606+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.514731+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:52.514731+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:53.490150+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: x(active, since 1.01005s) 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:53.490150+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: x(active, since 1.01005s) 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:54.153396+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.102:0/1916943896' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:54.153396+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.102:0/1916943896' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:54.442282+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.102:0/1165510801' entity='client.admin' cmd={"prefix": "config assimilate-conf"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:54.442282+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.102:0/1165510801' entity='client.admin' cmd={"prefix": "config assimilate-conf"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:54.494219+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e4: x(active, since 2s) 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:54.494219+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e4: x(active, since 2s) 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:54.801057+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "cephadm"} : dispatch 2026-03-21T06:48:25.443 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:54.801057+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "cephadm"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:55.502031+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:55.502031+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:55.504242+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e5: x(active, since 3s) 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:55.504242+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e5: x(active, since 3s) 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:55.836973+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.102:0/4113217554' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:55.836973+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.102:0/4113217554' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:58.527847+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:58.527847+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:58.528170+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:58.528170+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:58.533539+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:58.533539+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:58.533656+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e6: x(active, starting, since 0.00561459s) 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:58.533656+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e6: x(active, starting, since 0.00561459s) 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:58.535351+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:58.535351+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:58.536553+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:58.536553+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:58.537630+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:58.537630+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:58.537826+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:58.537826+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:58.537993+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:58.537993+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:58.546117+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:58.546117+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.047042+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.047042+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.051356+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.051356+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:47:59.051977+0000 mgr.x (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:47:59.051977+0000 mgr.x (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.054419+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.054419+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.060887+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.060887+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.064663+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.064663+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.067675+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.067675+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.072257+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.072257+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.074408+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.074408+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:25.444 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:59.537842+0000 mon.a (mon.0) 54 : cluster [DBG] mgrmap e7: x(active, since 1.0098s) 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:47:59.537842+0000 mon.a (mon.0) 54 : cluster [DBG] mgrmap e7: x(active, since 1.0098s) 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.538814+0000 mgr.x (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.538814+0000 mgr.x (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.543003+0000 mgr.x (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.543003+0000 mgr.x (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.908263+0000 mon.a (mon.0) 55 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "orchestrator"} : dispatch 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:47:59.908263+0000 mon.a (mon.0) 55 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "orchestrator"} : dispatch 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:00.100180+0000 mgr.x (mgr.14118) 4 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTING 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:00.100180+0000 mgr.x (mgr.14118) 4 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTING 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:00.201223+0000 mgr.x (mgr.14118) 5 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:00.201223+0000 mgr.x (mgr.14118) 5 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:00.308850+0000 mgr.x (mgr.14118) 6 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:00.308850+0000 mgr.x (mgr.14118) 6 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:00.308904+0000 mgr.x (mgr.14118) 7 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTED 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:00.308904+0000 mgr.x (mgr.14118) 7 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTED 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:00.309376+0000 mgr.x (mgr.14118) 8 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Client ('192.168.123.102', 32874) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:00.309376+0000 mgr.x (mgr.14118) 8 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Client ('192.168.123.102', 32874) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:00.309379+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:00.309379+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:00.537516+0000 mon.a (mon.0) 57 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "orchestrator"}]': finished 2026-03-21T06:48:25.445 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:00.537516+0000 mon.a (mon.0) 57 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "orchestrator"}]': finished 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:00.539788+0000 mon.a (mon.0) 58 : cluster [DBG] mgrmap e8: x(active, since 2s) 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:00.539788+0000 mon.a (mon.0) 58 : cluster [DBG] mgrmap e8: x(active, since 2s) 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:00.866761+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:00.866761+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:00.871909+0000 mon.a (mon.0) 60 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:00.871909+0000 mon.a (mon.0) 60 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:00.862870+0000 mgr.x (mgr.14118) 9 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:00.862870+0000 mgr.x (mgr.14118) 9 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:01.184829+0000 mgr.x (mgr.14118) 10 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:01.184829+0000 mgr.x (mgr.14118) 10 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:01.460422+0000 mgr.x (mgr.14118) 11 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:01.460422+0000 mgr.x (mgr.14118) 11 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:01.460668+0000 mgr.x (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:01.460668+0000 mgr.x (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:01.477702+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:01.477702+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:01.480261+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:01.480261+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:01.760765+0000 mgr.x (mgr.14118) 13 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:01.760765+0000 mgr.x (mgr.14118) 13 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:02.049007+0000 mgr.x (mgr.14118) 14 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:02.049007+0000 mgr.x (mgr.14118) 14 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:02.587117+0000 mgr.x (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:02.587117+0000 mgr.x (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:03.905211+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:03.905211+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:03.906295+0000 mon.a (mon.0) 64 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:03.906295+0000 mon.a (mon.0) 64 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:04.240830+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:04.240830+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:03.906013+0000 mgr.x (mgr.14118) 16 : cephadm [INF] Added host vm02 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:03.906013+0000 mgr.x (mgr.14118) 16 : cephadm [INF] Added host vm02 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:04.235767+0000 mgr.x (mgr.14118) 17 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:04.235767+0000 mgr.x (mgr.14118) 17 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:04.236801+0000 mgr.x (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:04.236801+0000 mgr.x (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:04.562737+0000 mgr.x (mgr.14118) 19 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:04.562737+0000 mgr.x (mgr.14118) 19 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:04.563606+0000 mgr.x (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:04.563606+0000 mgr.x (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:04.567739+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:04.567739+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:04.919608+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.102:0/3282338898' entity='client.admin' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:04.919608+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.102:0/3282338898' entity='client.admin' 2026-03-21T06:48:25.446 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:05.241068+0000 mon.a (mon.0) 68 : audit [INF] from='client.? 192.168.123.102:0/705651056' entity='client.admin' 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:05.241068+0000 mon.a (mon.0) 68 : audit [INF] from='client.? 192.168.123.102:0/705651056' entity='client.admin' 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:05.399776+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:05.399776+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:05.656724+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "dashboard"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:05.656724+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "dashboard"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:05.737887+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:05.737887+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:06.573803+0000 mon.a (mon.0) 72 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:06.573803+0000 mon.a (mon.0) 72 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:06.576279+0000 mon.a (mon.0) 73 : cluster [DBG] mgrmap e9: x(active, since 8s) 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:06.576279+0000 mon.a (mon.0) 73 : cluster [DBG] mgrmap e9: x(active, since 8s) 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:06.947047+0000 mon.a (mon.0) 74 : audit [DBG] from='client.? 192.168.123.102:0/2257204228' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:06.947047+0000 mon.a (mon.0) 74 : audit [DBG] from='client.? 192.168.123.102:0/2257204228' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:09.528899+0000 mon.a (mon.0) 75 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:09.528899+0000 mon.a (mon.0) 75 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:09.529177+0000 mon.a (mon.0) 76 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:09.529177+0000 mon.a (mon.0) 76 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:09.534390+0000 mon.a (mon.0) 77 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:09.534390+0000 mon.a (mon.0) 77 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:09.534484+0000 mon.a (mon.0) 78 : cluster [DBG] mgrmap e10: x(active, starting, since 0.00543034s) 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:09.534484+0000 mon.a (mon.0) 78 : cluster [DBG] mgrmap e10: x(active, starting, since 0.00543034s) 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.535036+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.535036+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.535833+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.535833+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.536634+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.536634+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.536754+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.536754+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.536847+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.536847+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:09.542468+0000 mon.a (mon.0) 84 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:09.542468+0000 mon.a (mon.0) 84 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.733700+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.733700+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.747375+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:25.447 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.747375+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.754178+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:09.754178+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:10.540431+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e11: x(active, since 1.01138s) 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:10.540431+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e11: x(active, since 1.01138s) 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:10.169504+0000 mgr.x (mgr.14152) 1 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTING 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:10.169504+0000 mgr.x (mgr.14152) 1 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTING 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:10.270812+0000 mgr.x (mgr.14152) 2 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:10.270812+0000 mgr.x (mgr.14152) 2 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:10.379369+0000 mgr.x (mgr.14152) 3 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:10.379369+0000 mgr.x (mgr.14152) 3 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:10.379416+0000 mgr.x (mgr.14152) 4 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTED 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:10.379416+0000 mgr.x (mgr.14152) 4 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTED 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:10.379868+0000 mgr.x (mgr.14152) 5 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Client ('192.168.123.102', 60300) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:10.379868+0000 mgr.x (mgr.14152) 5 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Client ('192.168.123.102', 60300) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:10.539699+0000 mgr.x (mgr.14152) 6 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:10.539699+0000 mgr.x (mgr.14152) 6 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:10.543919+0000 mgr.x (mgr.14152) 7 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:10.543919+0000 mgr.x (mgr.14152) 7 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:10.854917+0000 mgr.x (mgr.14152) 8 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "orch certmgr generate-certificates", "module_name": "dashboard", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:10.854917+0000 mgr.x (mgr.14152) 8 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "orch certmgr generate-certificates", "module_name": "dashboard", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:11.593890+0000 mon.a (mon.0) 89 : cluster [DBG] mgrmap e12: x(active, since 2s) 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:11.593890+0000 mon.a (mon.0) 89 : cluster [DBG] mgrmap e12: x(active, since 2s) 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.093702+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.093702+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.415641+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.415641+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.089751+0000 mgr.x (mgr.14152) 9 : audit [DBG] from='client.14166 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.089751+0000 mgr.x (mgr.14152) 9 : audit [DBG] from='client.14166 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.411813+0000 mgr.x (mgr.14152) 10 : audit [DBG] from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.411813+0000 mgr.x (mgr.14152) 10 : audit [DBG] from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.916726+0000 mon.a (mon.0) 92 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.916726+0000 mon.a (mon.0) 92 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:13.203820+0000 mon.a (mon.0) 93 : audit [DBG] from='client.? 192.168.123.102:0/2434461531' entity='client.admin' cmd={"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"} : dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:13.203820+0000 mon.a (mon.0) 93 : audit [DBG] from='client.? 192.168.123.102:0/2434461531' entity='client.admin' cmd={"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"} : dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:13.555919+0000 mon.a (mon.0) 94 : audit [INF] from='client.? 192.168.123.102:0/721893450' entity='client.admin' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:13.555919+0000 mon.a (mon.0) 94 : audit [INF] from='client.? 192.168.123.102:0/721893450' entity='client.admin' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.711292+0000 mgr.x (mgr.14152) 11 : audit [DBG] from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:12.711292+0000 mgr.x (mgr.14152) 11 : audit [DBG] from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:14.608592+0000 mon.a (mon.0) 95 : audit [INF] from='client.? 192.168.123.102:0/1702136911' entity='client.admin' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:14.608592+0000 mon.a (mon.0) 95 : audit [INF] from='client.? 192.168.123.102:0/1702136911' entity='client.admin' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:14.697746+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:14.697746+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:15.266639+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:15.266639+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:15.271954+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:15.271954+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:15.268547+0000 mgr.x (mgr.14152) 12 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:15.268547+0000 mgr.x (mgr.14152) 12 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:16.275985+0000 mon.a (mon.0) 99 : cluster [DBG] mgrmap e13: x(active, since 6s) 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:16.275985+0000 mon.a (mon.0) 99 : cluster [DBG] mgrmap e13: x(active, since 6s) 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.288184+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.288184+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.291087+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.291087+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.291911+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:25.448 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.291911+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.292731+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.292731+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.293263+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.293263+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.416070+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.416070+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.419237+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.419237+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.421755+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.421755+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.426463+0000 mon.a (mon.0) 108 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.426463+0000 mon.a (mon.0) 108 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.427256+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.427256+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.427695+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.427695+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.430299+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:16.430299+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:15.957062+0000 mgr.x (mgr.14152) 13 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:15.957062+0000 mgr.x (mgr.14152) 13 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:16.294027+0000 mgr.x (mgr.14152) 14 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:16.294027+0000 mgr.x (mgr.14152) 14 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:16.325888+0000 mgr.x (mgr.14152) 15 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:16.325888+0000 mgr.x (mgr.14152) 15 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:16.354553+0000 mgr.x (mgr.14152) 16 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:16.354553+0000 mgr.x (mgr.14152) 16 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:16.381714+0000 mgr.x (mgr.14152) 17 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:16.381714+0000 mgr.x (mgr.14152) 17 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:16.503567+0000 mgr.x (mgr.14152) 18 : cephadm [INF] Deploying cephadm binary to vm04 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:16.503567+0000 mgr.x (mgr.14152) 18 : cephadm [INF] Deploying cephadm binary to vm04 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:17.825992+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:17.825992+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:17.826396+0000 mgr.x (mgr.14152) 19 : cephadm [INF] Added host vm04 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:17.826396+0000 mgr.x (mgr.14152) 19 : cephadm [INF] Added host vm04 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:17.826677+0000 mon.a (mon.0) 113 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:17.826677+0000 mon.a (mon.0) 113 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:18.113003+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:18.113003+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:18.368400+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:18.368400+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:18.433860+0000 mgr.x (mgr.14152) 20 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:18.433860+0000 mgr.x (mgr.14152) 20 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:18.928895+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:18.928895+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:19.104814+0000 mgr.x (mgr.14152) 21 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm07", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:19.104814+0000 mgr.x (mgr.14152) 21 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm07", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:19.708613+0000 mgr.x (mgr.14152) 22 : cephadm [INF] Deploying cephadm binary to vm07 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:19.708613+0000 mgr.x (mgr.14152) 22 : cephadm [INF] Deploying cephadm binary to vm07 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.029193+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.029193+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.031967+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.031967+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.035357+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.035357+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.038086+0000 mon.a (mon.0) 120 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.038086+0000 mon.a (mon.0) 120 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.038646+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.038646+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.039370+0000 mon.a (mon.0) 122 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.039370+0000 mon.a (mon.0) 122 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.039897+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.039897+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:25.449 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:20.040563+0000 mgr.x (mgr.14152) 23 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:20.040563+0000 mgr.x (mgr.14152) 23 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:20.080611+0000 mgr.x (mgr.14152) 24 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:20.080611+0000 mgr.x (mgr.14152) 24 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:20.116690+0000 mgr.x (mgr.14152) 25 : cephadm [INF] Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:20.116690+0000 mgr.x (mgr.14152) 25 : cephadm [INF] Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:20.150701+0000 mgr.x (mgr.14152) 26 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:20.150701+0000 mgr.x (mgr.14152) 26 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.615400+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.615400+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.619283+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.619283+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.622655+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:20.622655+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:21.143323+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:21.143323+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:21.143828+0000 mgr.x (mgr.14152) 27 : cephadm [INF] Added host vm07 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:21.143828+0000 mgr.x (mgr.14152) 27 : cephadm [INF] Added host vm07 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:21.144202+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:21.144202+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:21.440241+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:21.440241+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:21.706038+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:21.706038+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:21.784850+0000 mgr.x (mgr.14152) 28 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:21.784850+0000 mgr.x (mgr.14152) 28 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:22.280754+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:22.280754+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:22.411837+0000 mon.a (mon.0) 132 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd={"prefix": "osd crush tunables", "profile": "default"} : dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:22.411837+0000 mon.a (mon.0) 132 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd={"prefix": "osd crush tunables", "profile": "default"} : dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.229669+0000 mon.a (mon.0) 133 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.229669+0000 mon.a (mon.0) 133 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.232858+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.232858+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.236368+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.236368+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.239045+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.239045+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.239534+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.239534+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.240204+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.240204+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.240646+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.240646+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.241301+0000 mgr.x (mgr.14152) 29 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.241301+0000 mgr.x (mgr.14152) 29 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.271217+0000 mgr.x (mgr.14152) 30 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.271217+0000 mgr.x (mgr.14152) 30 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.285120+0000 mon.a (mon.0) 140 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.285120+0000 mon.a (mon.0) 140 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:23.289349+0000 mon.a (mon.0) 141 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cluster 2026-03-21T06:48:23.289349+0000 mon.a (mon.0) 141 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.303461+0000 mgr.x (mgr.14152) 31 : cephadm [INF] Updating vm07:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.303461+0000 mgr.x (mgr.14152) 31 : cephadm [INF] Updating vm07:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.335772+0000 mgr.x (mgr.14152) 32 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.335772+0000 mgr.x (mgr.14152) 32 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.369557+0000 mon.a (mon.0) 142 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.369557+0000 mon.a (mon.0) 142 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.372382+0000 mon.a (mon.0) 143 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.372382+0000 mon.a (mon.0) 143 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.374780+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.374780+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.895775+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.895775+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.896325+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.896325+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.897294+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.897294+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.897802+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.897802+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.901129+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.901129+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.902098+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.902098+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.902507+0000 mon.a (mon.0) 151 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.902507+0000 mon.a (mon.0) 151 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.891168+0000 mgr.x (mgr.14152) 33 : audit [DBG] from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:23.891168+0000 mgr.x (mgr.14152) 33 : audit [DBG] from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.892375+0000 mgr.x (mgr.14152) 34 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c;count:3 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.892375+0000 mgr.x (mgr.14152) 34 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c;count:3 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.903006+0000 mgr.x (mgr.14152) 35 : cephadm [INF] Deploying daemon mon.c on vm07 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: cephadm 2026-03-21T06:48:23.903006+0000 mgr.x (mgr.14152) 35 : cephadm [INF] Deploying daemon mon.c on vm07 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:24.628820+0000 mon.a (mon.0) 152 : audit [DBG] from='client.? 192.168.123.107:0/3713273853' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:24.628820+0000 mon.a (mon.0) 152 : audit [DBG] from='client.? 192.168.123.107:0/3713273853' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:25.271212+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:25.271212+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:25.274482+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:25.274482+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:25.277002+0000 mon.a (mon.0) 155 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:25.277002+0000 mon.a (mon.0) 155 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:25.277419+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:25.277419+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:25.277895+0000 mon.a (mon.0) 157 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.451 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: audit 2026-03-21T06:48:25.277895+0000 mon.a (mon.0) 157 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:23.891168+0000 mgr.x (mgr.14152) 33 : audit [DBG] from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:23.891168+0000 mgr.x (mgr.14152) 33 : audit [DBG] from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.892375+0000 mgr.x (mgr.14152) 34 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c;count:3 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.892375+0000 mgr.x (mgr.14152) 34 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c;count:3 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.903006+0000 mgr.x (mgr.14152) 35 : cephadm [INF] Deploying daemon mon.c on vm07 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: cephadm 2026-03-21T06:48:23.903006+0000 mgr.x (mgr.14152) 35 : cephadm [INF] Deploying daemon mon.c on vm07 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:24.628820+0000 mon.a (mon.0) 152 : audit [DBG] from='client.? 192.168.123.107:0/3713273853' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:24.628820+0000 mon.a (mon.0) 152 : audit [DBG] from='client.? 192.168.123.107:0/3713273853' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:25.271212+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:25.271212+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:25.274482+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:25.274482+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:25.277002+0000 mon.a (mon.0) 155 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:25.277002+0000 mon.a (mon.0) 155 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:25.277419+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:25.277419+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:25.277895+0000 mon.a (mon.0) 157 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:25 vm02 bash[17657]: audit 2026-03-21T06:48:25.277895+0000 mon.a (mon.0) 157 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:25.818 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-21T06:48:25.819 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph mon dump -f json 2026-03-21T06:48:25.827 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:25 vm07 bash[19945]: debug 2026-03-21T06:48:25.437+0000 7f7b6a664640 1 mon.c@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-21T06:48:26.097 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:48:26.985 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.959+0000 7f1bda123d80 1 mon.b@-1(???) e0 preinit fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:26.985 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.983+0000 7f1bd2eb2640 0 mon.b@-1(synchronizing).mds e1 new map 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.987+0000 7f1bd2eb2640 0 mon.b@-1(synchronizing).mds e1 print_map 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: e1 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: btime 2026-03-21T06:47:47:923913+0000 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: enable_multiple, ever_enabled_multiple: 1,1 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: default compat: compat={},rocompat={},incompat={1=base v0.20,2=client writeable ranges,3=default file layouts on dirs,4=dir inode in separate object,5=mds uses versioned encoding,6=dirfrag is stored in omap,8=no anchor table,9=file layout v2,10=snaprealm v2,11=minor log segments,12=quiesce subvolumes} 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: legacy client fscid: -1 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: No filesystems configured 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.987+0000 7f1bd2eb2640 1 mon.b@-1(synchronizing).osd e0 _set_cache_ratios kv ratio 0.25 inc ratio 0.375 full ratio 0.375 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.987+0000 7f1bd2eb2640 1 mon.b@-1(synchronizing).osd e0 register_cache_with_pcm pcm target: 2147483648 pcm max: 1020054732 pcm min: 134217728 inc_osd_cache size: 1 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.987+0000 7f1bd2eb2640 1 mon.b@-1(synchronizing).osd e1 e1: 0 total, 0 up, 0 in 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.987+0000 7f1bd2eb2640 1 mon.b@-1(synchronizing).osd e2 e2: 0 total, 0 up, 0 in 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.987+0000 7f1bd2eb2640 1 mon.b@-1(synchronizing).osd e3 e3: 0 total, 0 up, 0 in 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.987+0000 7f1bd2eb2640 1 mon.b@-1(synchronizing).osd e4 e4: 0 total, 0 up, 0 in 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.987+0000 7f1bd2eb2640 0 mon.b@-1(synchronizing).osd e4 crush map has features 3314932999778484224, adjusting msgr requires 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.987+0000 7f1bd2eb2640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.987+0000 7f1bd2eb2640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: debug 2026-03-21T06:48:26.987+0000 7f1bd2eb2640 0 mon.b@-1(synchronizing).osd e4 crush map has features 288514050185494528, adjusting msgr requires 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:47.924339+0000 mon.a (mon.0) 0 : cluster [INF] mkfs b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:47.924339+0000 mon.a (mon.0) 0 : cluster [INF] mkfs b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:47.919314+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:47.919314+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920244+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920244+0000 mon.a (mon.0) 1 : cluster [INF] mon.a is new leader, mons a in quorum (ranks 0) 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920278+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920278+0000 mon.a (mon.0) 2 : cluster [DBG] monmap epoch 1 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920283+0000 mon.a (mon.0) 3 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920283+0000 mon.a (mon.0) 3 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920287+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920287+0000 mon.a (mon.0) 4 : cluster [DBG] last_changed 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920297+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920297+0000 mon.a (mon.0) 5 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920302+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920302+0000 mon.a (mon.0) 6 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:27.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920311+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920311+0000 mon.a (mon.0) 7 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920314+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920314+0000 mon.a (mon.0) 8 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920633+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920633+0000 mon.a (mon.0) 9 : cluster [DBG] fsmap 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920648+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.920648+0000 mon.a (mon.0) 10 : cluster [DBG] osdmap e1: 0 total, 0 up, 0 in 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.921256+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:48.921256+0000 mon.a (mon.0) 11 : cluster [DBG] mgrmap e1: no daemons active 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:48.990782+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1389498430' entity='client.admin' 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:48.990782+0000 mon.a (mon.0) 12 : audit [INF] from='client.? 192.168.123.102:0/1389498430' entity='client.admin' 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:49.586159+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/4046014269' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:49.586159+0000 mon.a (mon.0) 13 : audit [DBG] from='client.? 192.168.123.102:0/4046014269' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:51.810831+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/788823380' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:51.810831+0000 mon.a (mon.0) 14 : audit [DBG] from='client.? 192.168.123.102:0/788823380' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:52.480256+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:52.480256+0000 mon.a (mon.0) 15 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:52.484714+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: x(active, starting, since 0.00461082s) 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:52.484714+0000 mon.a (mon.0) 16 : cluster [DBG] mgrmap e2: x(active, starting, since 0.00461082s) 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.486227+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.486227+0000 mon.a (mon.0) 17 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.486600+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.486600+0000 mon.a (mon.0) 18 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.486956+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.486956+0000 mon.a (mon.0) 19 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.487347+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.487347+0000 mon.a (mon.0) 20 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.487688+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.487688+0000 mon.a (mon.0) 21 : audit [DBG] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:52.494572+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: cluster 2026-03-21T06:47:52.494572+0000 mon.a (mon.0) 22 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.503203+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.503203+0000 mon.a (mon.0) 23 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.507496+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.507496+0000 mon.a (mon.0) 24 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.507813+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:26 vm04 bash[20194]: audit 2026-03-21T06:47:52.507813+0000 mon.a (mon.0) 25 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:52.511606+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:52.511606+0000 mon.a (mon.0) 26 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:52.514731+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:52.514731+0000 mon.a (mon.0) 27 : audit [INF] from='mgr.14100 192.168.123.102:0/1954665080' entity='mgr.x' 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:53.490150+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: x(active, since 1.01005s) 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:53.490150+0000 mon.a (mon.0) 28 : cluster [DBG] mgrmap e3: x(active, since 1.01005s) 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:54.153396+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.102:0/1916943896' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:54.153396+0000 mon.a (mon.0) 29 : audit [DBG] from='client.? 192.168.123.102:0/1916943896' entity='client.admin' cmd={"prefix": "status", "format": "json-pretty"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:54.442282+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.102:0/1165510801' entity='client.admin' cmd={"prefix": "config assimilate-conf"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:54.442282+0000 mon.a (mon.0) 30 : audit [INF] from='client.? 192.168.123.102:0/1165510801' entity='client.admin' cmd={"prefix": "config assimilate-conf"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:54.494219+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e4: x(active, since 2s) 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:54.494219+0000 mon.a (mon.0) 31 : cluster [DBG] mgrmap e4: x(active, since 2s) 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:54.801057+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "cephadm"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:54.801057+0000 mon.a (mon.0) 32 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "cephadm"} : dispatch 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:55.502031+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:55.502031+0000 mon.a (mon.0) 33 : audit [INF] from='client.? 192.168.123.102:0/2441753982' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "cephadm"}]': finished 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:55.504242+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e5: x(active, since 3s) 2026-03-21T06:48:27.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:55.504242+0000 mon.a (mon.0) 34 : cluster [DBG] mgrmap e5: x(active, since 3s) 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:55.836973+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.102:0/4113217554' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:55.836973+0000 mon.a (mon.0) 35 : audit [DBG] from='client.? 192.168.123.102:0/4113217554' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:58.527847+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:58.527847+0000 mon.a (mon.0) 36 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:58.528170+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:58.528170+0000 mon.a (mon.0) 37 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:58.533539+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:58.533539+0000 mon.a (mon.0) 38 : cluster [DBG] osdmap e2: 0 total, 0 up, 0 in 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:58.533656+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e6: x(active, starting, since 0.00561459s) 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:58.533656+0000 mon.a (mon.0) 39 : cluster [DBG] mgrmap e6: x(active, starting, since 0.00561459s) 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:58.535351+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:58.535351+0000 mon.a (mon.0) 40 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:58.536553+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:58.536553+0000 mon.a (mon.0) 41 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:58.537630+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:58.537630+0000 mon.a (mon.0) 42 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:58.537826+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:58.537826+0000 mon.a (mon.0) 43 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:58.537993+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:58.537993+0000 mon.a (mon.0) 44 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:58.546117+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:58.546117+0000 mon.a (mon.0) 45 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.047042+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.047042+0000 mon.a (mon.0) 46 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.051356+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.051356+0000 mon.a (mon.0) 47 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:47:59.051977+0000 mgr.x (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:47:59.051977+0000 mgr.x (mgr.14118) 1 : cephadm [INF] Found migration_current of "None". Setting to last migration. 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.054419+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.054419+0000 mon.a (mon.0) 48 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.060887+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.060887+0000 mon.a (mon.0) 49 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.064663+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.064663+0000 mon.a (mon.0) 50 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.067675+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.067675+0000 mon.a (mon.0) 51 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.072257+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.072257+0000 mon.a (mon.0) 52 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.074408+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.074408+0000 mon.a (mon.0) 53 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:59.537842+0000 mon.a (mon.0) 54 : cluster [DBG] mgrmap e7: x(active, since 1.0098s) 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:47:59.537842+0000 mon.a (mon.0) 54 : cluster [DBG] mgrmap e7: x(active, since 1.0098s) 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.538814+0000 mgr.x (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.538814+0000 mgr.x (mgr.14118) 2 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.543003+0000 mgr.x (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.543003+0000 mgr.x (mgr.14118) 3 : audit [DBG] from='client.14122 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.908263+0000 mon.a (mon.0) 55 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "orchestrator"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:47:59.908263+0000 mon.a (mon.0) 55 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "orchestrator"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:00.100180+0000 mgr.x (mgr.14118) 4 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTING 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:00.100180+0000 mgr.x (mgr.14118) 4 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTING 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:00.201223+0000 mgr.x (mgr.14118) 5 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:00.201223+0000 mgr.x (mgr.14118) 5 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:00.308850+0000 mgr.x (mgr.14118) 6 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:00.308850+0000 mgr.x (mgr.14118) 6 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:00.308904+0000 mgr.x (mgr.14118) 7 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTED 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:00.308904+0000 mgr.x (mgr.14118) 7 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Bus STARTED 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:00.309376+0000 mgr.x (mgr.14118) 8 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Client ('192.168.123.102', 32874) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:00.309376+0000 mgr.x (mgr.14118) 8 : cephadm [INF] [21/Mar/2026:06:48:00] ENGINE Client ('192.168.123.102', 32874) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:00.309379+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:00.309379+0000 mon.a (mon.0) 56 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:00.537516+0000 mon.a (mon.0) 57 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "orchestrator"}]': finished 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:00.537516+0000 mon.a (mon.0) 57 : audit [INF] from='client.? 192.168.123.102:0/4010957210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "orchestrator"}]': finished 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:00.539788+0000 mon.a (mon.0) 58 : cluster [DBG] mgrmap e8: x(active, since 2s) 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:00.539788+0000 mon.a (mon.0) 58 : cluster [DBG] mgrmap e8: x(active, since 2s) 2026-03-21T06:48:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:00.866761+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:00.866761+0000 mon.a (mon.0) 59 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:00.871909+0000 mon.a (mon.0) 60 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:00.871909+0000 mon.a (mon.0) 60 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:00.862870+0000 mgr.x (mgr.14118) 9 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:00.862870+0000 mgr.x (mgr.14118) 9 : audit [DBG] from='client.14132 -' entity='client.admin' cmd=[{"prefix": "orch set backend", "module_name": "cephadm", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:01.184829+0000 mgr.x (mgr.14118) 10 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:01.184829+0000 mgr.x (mgr.14118) 10 : audit [DBG] from='client.14134 -' entity='client.admin' cmd=[{"prefix": "cephadm set-user", "user": "root", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:01.460422+0000 mgr.x (mgr.14118) 11 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:01.460422+0000 mgr.x (mgr.14118) 11 : audit [DBG] from='client.14136 -' entity='client.admin' cmd=[{"prefix": "cephadm generate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:01.460668+0000 mgr.x (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:01.460668+0000 mgr.x (mgr.14118) 12 : cephadm [INF] Generating ssh key... 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:01.477702+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:01.477702+0000 mon.a (mon.0) 61 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:01.480261+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:01.480261+0000 mon.a (mon.0) 62 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:01.760765+0000 mgr.x (mgr.14118) 13 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:01.760765+0000 mgr.x (mgr.14118) 13 : audit [DBG] from='client.14138 -' entity='client.admin' cmd=[{"prefix": "cephadm get-pub-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:02.049007+0000 mgr.x (mgr.14118) 14 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:02.049007+0000 mgr.x (mgr.14118) 14 : audit [DBG] from='client.14140 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm02", "addr": "192.168.123.102", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:02.587117+0000 mgr.x (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:02.587117+0000 mgr.x (mgr.14118) 15 : cephadm [INF] Deploying cephadm binary to vm02 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:03.905211+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:03.905211+0000 mon.a (mon.0) 63 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:03.906295+0000 mon.a (mon.0) 64 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:03.906295+0000 mon.a (mon.0) 64 : audit [DBG] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:04.240830+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:04.240830+0000 mon.a (mon.0) 65 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:03.906013+0000 mgr.x (mgr.14118) 16 : cephadm [INF] Added host vm02 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:03.906013+0000 mgr.x (mgr.14118) 16 : cephadm [INF] Added host vm02 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:04.235767+0000 mgr.x (mgr.14118) 17 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:04.235767+0000 mgr.x (mgr.14118) 17 : audit [DBG] from='client.14142 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:04.236801+0000 mgr.x (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:04.236801+0000 mgr.x (mgr.14118) 18 : cephadm [INF] Saving service mon spec with placement count:5 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:04.562737+0000 mgr.x (mgr.14118) 19 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:04.562737+0000 mgr.x (mgr.14118) 19 : audit [DBG] from='client.14144 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "unmanaged": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:04.563606+0000 mgr.x (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:04.563606+0000 mgr.x (mgr.14118) 20 : cephadm [INF] Saving service mgr spec with placement count:2 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:04.567739+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:04.567739+0000 mon.a (mon.0) 66 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:04.919608+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.102:0/3282338898' entity='client.admin' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:04.919608+0000 mon.a (mon.0) 67 : audit [INF] from='client.? 192.168.123.102:0/3282338898' entity='client.admin' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:05.241068+0000 mon.a (mon.0) 68 : audit [INF] from='client.? 192.168.123.102:0/705651056' entity='client.admin' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:05.241068+0000 mon.a (mon.0) 68 : audit [INF] from='client.? 192.168.123.102:0/705651056' entity='client.admin' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:05.399776+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:05.399776+0000 mon.a (mon.0) 69 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:05.656724+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "dashboard"} : dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:05.656724+0000 mon.a (mon.0) 70 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd={"prefix": "mgr module enable", "module": "dashboard"} : dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:05.737887+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:05.737887+0000 mon.a (mon.0) 71 : audit [INF] from='mgr.14118 192.168.123.102:0/699886669' entity='mgr.x' 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:06.573803+0000 mon.a (mon.0) 72 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:06.573803+0000 mon.a (mon.0) 72 : audit [INF] from='client.? 192.168.123.102:0/1072319210' entity='client.admin' cmd='[{"prefix": "mgr module enable", "module": "dashboard"}]': finished 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:06.576279+0000 mon.a (mon.0) 73 : cluster [DBG] mgrmap e9: x(active, since 8s) 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:06.576279+0000 mon.a (mon.0) 73 : cluster [DBG] mgrmap e9: x(active, since 8s) 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:06.947047+0000 mon.a (mon.0) 74 : audit [DBG] from='client.? 192.168.123.102:0/2257204228' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:06.947047+0000 mon.a (mon.0) 74 : audit [DBG] from='client.? 192.168.123.102:0/2257204228' entity='client.admin' cmd={"prefix": "mgr stat"} : dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:09.528899+0000 mon.a (mon.0) 75 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:09.528899+0000 mon.a (mon.0) 75 : cluster [INF] Active manager daemon x restarted 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:09.529177+0000 mon.a (mon.0) 76 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:09.529177+0000 mon.a (mon.0) 76 : cluster [INF] Activating manager daemon x 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:09.534390+0000 mon.a (mon.0) 77 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:09.534390+0000 mon.a (mon.0) 77 : cluster [DBG] osdmap e3: 0 total, 0 up, 0 in 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:09.534484+0000 mon.a (mon.0) 78 : cluster [DBG] mgrmap e10: x(active, starting, since 0.00543034s) 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:09.534484+0000 mon.a (mon.0) 78 : cluster [DBG] mgrmap e10: x(active, starting, since 0.00543034s) 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.535036+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.535036+0000 mon.a (mon.0) 79 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.535833+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.535833+0000 mon.a (mon.0) 80 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr metadata", "who": "x", "id": "x"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.536634+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.536634+0000 mon.a (mon.0) 81 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mds metadata"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.536754+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.536754+0000 mon.a (mon.0) 82 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.536847+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.536847+0000 mon.a (mon.0) 83 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:09.542468+0000 mon.a (mon.0) 84 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:09.542468+0000 mon.a (mon.0) 84 : cluster [INF] Manager daemon x is now available 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.733700+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.733700+0000 mon.a (mon.0) 85 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.747375+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.747375+0000 mon.a (mon.0) 86 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/mirror_snapshot_schedule"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.754178+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:09.754178+0000 mon.a (mon.0) 87 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix":"config rm","who":"mgr","name":"mgr/rbd_support/x/trash_purge_schedule"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:10.540431+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e11: x(active, since 1.01138s) 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:10.540431+0000 mon.a (mon.0) 88 : cluster [DBG] mgrmap e11: x(active, since 1.01138s) 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:10.169504+0000 mgr.x (mgr.14152) 1 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTING 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:10.169504+0000 mgr.x (mgr.14152) 1 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTING 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:10.270812+0000 mgr.x (mgr.14152) 2 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:10.270812+0000 mgr.x (mgr.14152) 2 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on http://192.168.123.102:8765 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:10.379369+0000 mgr.x (mgr.14152) 3 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:10.379369+0000 mgr.x (mgr.14152) 3 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Serving on https://192.168.123.102:7150 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:10.379416+0000 mgr.x (mgr.14152) 4 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTED 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:10.379416+0000 mgr.x (mgr.14152) 4 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Bus STARTED 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:10.379868+0000 mgr.x (mgr.14152) 5 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Client ('192.168.123.102', 60300) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:10.379868+0000 mgr.x (mgr.14152) 5 : cephadm [INF] [21/Mar/2026:06:48:10] ENGINE Client ('192.168.123.102', 60300) lost — peer dropped the TLS connection suddenly, during handshake: (6, 'TLS/SSL connection has been closed (EOF) (_ssl.c:1147)') 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:10.539699+0000 mgr.x (mgr.14152) 6 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:10.539699+0000 mgr.x (mgr.14152) 6 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "get_command_descriptions"}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:10.543919+0000 mgr.x (mgr.14152) 7 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:10.543919+0000 mgr.x (mgr.14152) 7 : audit [DBG] from='client.14156 -' entity='client.admin' cmd=[{"prefix": "mgr_status"}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:10.854917+0000 mgr.x (mgr.14152) 8 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "orch certmgr generate-certificates", "module_name": "dashboard", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:10.854917+0000 mgr.x (mgr.14152) 8 : audit [DBG] from='client.14164 -' entity='client.admin' cmd=[{"prefix": "orch certmgr generate-certificates", "module_name": "dashboard", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:11.593890+0000 mon.a (mon.0) 89 : cluster [DBG] mgrmap e12: x(active, since 2s) 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:11.593890+0000 mon.a (mon.0) 89 : cluster [DBG] mgrmap e12: x(active, since 2s) 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.093702+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.093702+0000 mon.a (mon.0) 90 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.415641+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.415641+0000 mon.a (mon.0) 91 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.089751+0000 mgr.x (mgr.14152) 9 : audit [DBG] from='client.14166 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.089751+0000 mgr.x (mgr.14152) 9 : audit [DBG] from='client.14166 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.411813+0000 mgr.x (mgr.14152) 10 : audit [DBG] from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.411813+0000 mgr.x (mgr.14152) 10 : audit [DBG] from='client.14168 -' entity='client.admin' cmd=[{"prefix": "dashboard set-ssl-certificate-key", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.916726+0000 mon.a (mon.0) 92 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.916726+0000 mon.a (mon.0) 92 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:13.203820+0000 mon.a (mon.0) 93 : audit [DBG] from='client.? 192.168.123.102:0/2434461531' entity='client.admin' cmd={"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:13.203820+0000 mon.a (mon.0) 93 : audit [DBG] from='client.? 192.168.123.102:0/2434461531' entity='client.admin' cmd={"prefix": "config get", "who": "mgr", "key": "mgr/dashboard/ssl_server_port"} : dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:13.555919+0000 mon.a (mon.0) 94 : audit [INF] from='client.? 192.168.123.102:0/721893450' entity='client.admin' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:13.555919+0000 mon.a (mon.0) 94 : audit [INF] from='client.? 192.168.123.102:0/721893450' entity='client.admin' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.711292+0000 mgr.x (mgr.14152) 11 : audit [DBG] from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:12.711292+0000 mgr.x (mgr.14152) 11 : audit [DBG] from='client.14170 -' entity='client.admin' cmd=[{"prefix": "dashboard ac-user-create", "username": "admin", "rolename": "administrator", "force_password": true, "pwd_update_required": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:14.608592+0000 mon.a (mon.0) 95 : audit [INF] from='client.? 192.168.123.102:0/1702136911' entity='client.admin' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:14.608592+0000 mon.a (mon.0) 95 : audit [INF] from='client.? 192.168.123.102:0/1702136911' entity='client.admin' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:14.697746+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:14.697746+0000 mon.a (mon.0) 96 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:15.266639+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:15.266639+0000 mon.a (mon.0) 97 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:15.271954+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:15.271954+0000 mon.a (mon.0) 98 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:15.268547+0000 mgr.x (mgr.14152) 12 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:15.268547+0000 mgr.x (mgr.14152) 12 : audit [DBG] from='client.14178 -' entity='client.admin' cmd=[{"prefix": "orch client-keyring set", "entity": "client.admin", "placement": "*", "mode": "0755", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:16.275985+0000 mon.a (mon.0) 99 : cluster [DBG] mgrmap e13: x(active, since 6s) 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:16.275985+0000 mon.a (mon.0) 99 : cluster [DBG] mgrmap e13: x(active, since 6s) 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.288184+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.288184+0000 mon.a (mon.0) 100 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.261 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.291087+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.291087+0000 mon.a (mon.0) 101 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.291911+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.291911+0000 mon.a (mon.0) 102 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm02", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.292731+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.292731+0000 mon.a (mon.0) 103 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.293263+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.293263+0000 mon.a (mon.0) 104 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.416070+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.416070+0000 mon.a (mon.0) 105 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.419237+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.419237+0000 mon.a (mon.0) 106 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.421755+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.421755+0000 mon.a (mon.0) 107 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.426463+0000 mon.a (mon.0) 108 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.426463+0000 mon.a (mon.0) 108 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.427256+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.427256+0000 mon.a (mon.0) 109 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.427695+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.427695+0000 mon.a (mon.0) 110 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.430299+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:16.430299+0000 mon.a (mon.0) 111 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:15.957062+0000 mgr.x (mgr.14152) 13 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:15.957062+0000 mgr.x (mgr.14152) 13 : audit [DBG] from='client.14180 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm04", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:16.294027+0000 mgr.x (mgr.14152) 14 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:16.294027+0000 mgr.x (mgr.14152) 14 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:16.325888+0000 mgr.x (mgr.14152) 15 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:16.325888+0000 mgr.x (mgr.14152) 15 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:16.354553+0000 mgr.x (mgr.14152) 16 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:16.354553+0000 mgr.x (mgr.14152) 16 : cephadm [INF] Updating vm02:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:16.381714+0000 mgr.x (mgr.14152) 17 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:16.381714+0000 mgr.x (mgr.14152) 17 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:16.503567+0000 mgr.x (mgr.14152) 18 : cephadm [INF] Deploying cephadm binary to vm04 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:16.503567+0000 mgr.x (mgr.14152) 18 : cephadm [INF] Deploying cephadm binary to vm04 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:17.825992+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:17.825992+0000 mon.a (mon.0) 112 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:17.826396+0000 mgr.x (mgr.14152) 19 : cephadm [INF] Added host vm04 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:17.826396+0000 mgr.x (mgr.14152) 19 : cephadm [INF] Added host vm04 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:17.826677+0000 mon.a (mon.0) 113 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:17.826677+0000 mon.a (mon.0) 113 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:18.113003+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:18.113003+0000 mon.a (mon.0) 114 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:18.368400+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:18.368400+0000 mon.a (mon.0) 115 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:18.433860+0000 mgr.x (mgr.14152) 20 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:18.433860+0000 mgr.x (mgr.14152) 20 : audit [DBG] from='client.14182 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:18.928895+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:18.928895+0000 mon.a (mon.0) 116 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:19.104814+0000 mgr.x (mgr.14152) 21 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm07", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:19.104814+0000 mgr.x (mgr.14152) 21 : audit [DBG] from='client.14184 -' entity='client.admin' cmd=[{"prefix": "orch host add", "hostname": "vm07", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:19.708613+0000 mgr.x (mgr.14152) 22 : cephadm [INF] Deploying cephadm binary to vm07 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:19.708613+0000 mgr.x (mgr.14152) 22 : cephadm [INF] Deploying cephadm binary to vm07 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.029193+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.029193+0000 mon.a (mon.0) 117 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.031967+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.031967+0000 mon.a (mon.0) 118 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.035357+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.035357+0000 mon.a (mon.0) 119 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.038086+0000 mon.a (mon.0) 120 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.038086+0000 mon.a (mon.0) 120 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.038646+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.038646+0000 mon.a (mon.0) 121 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm04", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.039370+0000 mon.a (mon.0) 122 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.039370+0000 mon.a (mon.0) 122 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.262 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.039897+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.039897+0000 mon.a (mon.0) 123 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:20.040563+0000 mgr.x (mgr.14152) 23 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:20.040563+0000 mgr.x (mgr.14152) 23 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:20.080611+0000 mgr.x (mgr.14152) 24 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:20.080611+0000 mgr.x (mgr.14152) 24 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:20.116690+0000 mgr.x (mgr.14152) 25 : cephadm [INF] Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:20.116690+0000 mgr.x (mgr.14152) 25 : cephadm [INF] Updating vm04:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:20.150701+0000 mgr.x (mgr.14152) 26 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:20.150701+0000 mgr.x (mgr.14152) 26 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.615400+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.615400+0000 mon.a (mon.0) 124 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.619283+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.619283+0000 mon.a (mon.0) 125 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.622655+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:20.622655+0000 mon.a (mon.0) 126 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:21.143323+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:21.143323+0000 mon.a (mon.0) 127 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:21.143828+0000 mgr.x (mgr.14152) 27 : cephadm [INF] Added host vm07 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:21.143828+0000 mgr.x (mgr.14152) 27 : cephadm [INF] Added host vm07 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:21.144202+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:21.144202+0000 mon.a (mon.0) 128 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:21.440241+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:21.440241+0000 mon.a (mon.0) 129 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:21.706038+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:21.706038+0000 mon.a (mon.0) 130 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:21.784850+0000 mgr.x (mgr.14152) 28 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:21.784850+0000 mgr.x (mgr.14152) 28 : audit [DBG] from='client.14186 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:22.280754+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:22.280754+0000 mon.a (mon.0) 131 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:22.411837+0000 mon.a (mon.0) 132 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd={"prefix": "osd crush tunables", "profile": "default"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:22.411837+0000 mon.a (mon.0) 132 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd={"prefix": "osd crush tunables", "profile": "default"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.229669+0000 mon.a (mon.0) 133 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.229669+0000 mon.a (mon.0) 133 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.232858+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.232858+0000 mon.a (mon.0) 134 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.236368+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.236368+0000 mon.a (mon.0) 135 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.239045+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.239045+0000 mon.a (mon.0) 136 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.239534+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.239534+0000 mon.a (mon.0) 137 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd/host:vm07", "name": "osd_memory_target"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.240204+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.240204+0000 mon.a (mon.0) 138 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.240646+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.240646+0000 mon.a (mon.0) 139 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.241301+0000 mgr.x (mgr.14152) 29 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.241301+0000 mgr.x (mgr.14152) 29 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.271217+0000 mgr.x (mgr.14152) 30 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:27.263 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.271217+0000 mgr.x (mgr.14152) 30 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.285120+0000 mon.a (mon.0) 140 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.285120+0000 mon.a (mon.0) 140 : audit [INF] from='client.? 192.168.123.102:0/3808115572' entity='client.admin' cmd='[{"prefix": "osd crush tunables", "profile": "default"}]': finished 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:23.289349+0000 mon.a (mon.0) 141 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cluster 2026-03-21T06:48:23.289349+0000 mon.a (mon.0) 141 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.303461+0000 mgr.x (mgr.14152) 31 : cephadm [INF] Updating vm07:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.303461+0000 mgr.x (mgr.14152) 31 : cephadm [INF] Updating vm07:/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.335772+0000 mgr.x (mgr.14152) 32 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.335772+0000 mgr.x (mgr.14152) 32 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.client.admin.keyring 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.369557+0000 mon.a (mon.0) 142 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.369557+0000 mon.a (mon.0) 142 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.372382+0000 mon.a (mon.0) 143 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.372382+0000 mon.a (mon.0) 143 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.374780+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.374780+0000 mon.a (mon.0) 144 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.895775+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.895775+0000 mon.a (mon.0) 145 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.896325+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.896325+0000 mon.a (mon.0) 146 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.897294+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.897294+0000 mon.a (mon.0) 147 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.897802+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.897802+0000 mon.a (mon.0) 148 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.901129+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.901129+0000 mon.a (mon.0) 149 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.902098+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.902098+0000 mon.a (mon.0) 150 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.902507+0000 mon.a (mon.0) 151 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.902507+0000 mon.a (mon.0) 151 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.891168+0000 mgr.x (mgr.14152) 33 : audit [DBG] from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:23.891168+0000 mgr.x (mgr.14152) 33 : audit [DBG] from='client.14190 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mon", "placement": "3;vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.892375+0000 mgr.x (mgr.14152) 34 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c;count:3 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.892375+0000 mgr.x (mgr.14152) 34 : cephadm [INF] Saving service mon spec with placement vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c;count:3 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.903006+0000 mgr.x (mgr.14152) 35 : cephadm [INF] Deploying daemon mon.c on vm07 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: cephadm 2026-03-21T06:48:23.903006+0000 mgr.x (mgr.14152) 35 : cephadm [INF] Deploying daemon mon.c on vm07 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:24.628820+0000 mon.a (mon.0) 152 : audit [DBG] from='client.? 192.168.123.107:0/3713273853' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:24.628820+0000 mon.a (mon.0) 152 : audit [DBG] from='client.? 192.168.123.107:0/3713273853' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:25.271212+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:25.271212+0000 mon.a (mon.0) 153 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:25.274482+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:25.274482+0000 mon.a (mon.0) 154 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:25.277002+0000 mon.a (mon.0) 155 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:25.277002+0000 mon.a (mon.0) 155 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:25.277419+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:25.277419+0000 mon.a (mon.0) 156 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:25.277895+0000 mon.a (mon.0) 157 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: audit 2026-03-21T06:48:25.277895+0000 mon.a (mon.0) 157 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:27.264 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:27 vm04 bash[20194]: debug 2026-03-21T06:48:27.007+0000 7f1bd2eb2640 1 mon.b@-1(synchronizing).paxosservice(auth 1..3) refresh upgraded, format 0 -> 3 2026-03-21T06:48:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cephadm 2026-03-21T06:48:25.278385+0000 mgr.x (mgr.14152) 36 : cephadm [INF] Deploying daemon mon.b on vm04 2026-03-21T06:48:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cephadm 2026-03-21T06:48:25.278385+0000 mgr.x (mgr.14152) 36 : cephadm [INF] Deploying daemon mon.b on vm04 2026-03-21T06:48:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:25.441939+0000 mon.a (mon.0) 159 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:25.441939+0000 mon.a (mon.0) 159 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:25.442139+0000 mon.a (mon.0) 160 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:25.442139+0000 mon.a (mon.0) 160 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:25.442253+0000 mon.a (mon.0) 161 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:25.442253+0000 mon.a (mon.0) 161 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:26.438924+0000 mon.a (mon.0) 162 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:26.438924+0000 mon.a (mon.0) 162 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:27.005379+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:27.005379+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:27.439344+0000 mon.a (mon.0) 164 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:27.439344+0000 mon.a (mon.0) 164 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:27.444989+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:27.444989+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:28.004918+0000 mon.a (mon.0) 165 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:28.004918+0000 mon.a (mon.0) 165 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:28.438505+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:28.438505+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:29.005037+0000 mon.a (mon.0) 167 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:29.005037+0000 mon.a (mon.0) 167 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:29.438371+0000 mon.a (mon.0) 168 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:29.438371+0000 mon.a (mon.0) 168 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:29.537781+0000 mgr.x (mgr.14152) 37 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:29.537781+0000 mgr.x (mgr.14152) 37 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.005157+0000 mon.a (mon.0) 169 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.005157+0000 mon.a (mon.0) 169 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.438749+0000 mon.a (mon.0) 170 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.438749+0000 mon.a (mon.0) 170 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.446776+0000 mon.a (mon.0) 171 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.446776+0000 mon.a (mon.0) 171 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450429+0000 mon.a (mon.0) 172 : cluster [DBG] monmap epoch 2 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450429+0000 mon.a (mon.0) 172 : cluster [DBG] monmap epoch 2 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450437+0000 mon.a (mon.0) 173 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450437+0000 mon.a (mon.0) 173 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450441+0000 mon.a (mon.0) 174 : cluster [DBG] last_changed 2026-03-21T06:48:25.438317+0000 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450441+0000 mon.a (mon.0) 174 : cluster [DBG] last_changed 2026-03-21T06:48:25.438317+0000 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450445+0000 mon.a (mon.0) 175 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450445+0000 mon.a (mon.0) 175 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450448+0000 mon.a (mon.0) 176 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450448+0000 mon.a (mon.0) 176 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450452+0000 mon.a (mon.0) 177 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450452+0000 mon.a (mon.0) 177 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450455+0000 mon.a (mon.0) 178 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450455+0000 mon.a (mon.0) 178 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450458+0000 mon.a (mon.0) 179 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450458+0000 mon.a (mon.0) 179 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450786+0000 mon.a (mon.0) 180 : cluster [DBG] fsmap 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450786+0000 mon.a (mon.0) 180 : cluster [DBG] fsmap 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450802+0000 mon.a (mon.0) 181 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450802+0000 mon.a (mon.0) 181 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:30.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450935+0000 mon.a (mon.0) 182 : cluster [DBG] mgrmap e13: x(active, since 20s) 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.450935+0000 mon.a (mon.0) 182 : cluster [DBG] mgrmap e13: x(active, since 20s) 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.451013+0000 mon.a (mon.0) 183 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: cluster 2026-03-21T06:48:30.451013+0000 mon.a (mon.0) 183 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.454382+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.454382+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.460349+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.460349+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.464229+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.464229+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.467207+0000 mon.a (mon.0) 187 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.467207+0000 mon.a (mon.0) 187 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.480753+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:30.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:30 vm02 bash[17657]: audit 2026-03-21T06:48:30.480753+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:30.899 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:48:30.899 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":2,"fsid":"b16ecafc-24f1-11f1-8ede-8330751617ee","modified":"2026-03-21T06:48:25.438317Z","created":"2026-03-21T06:47:46.690400Z","min_mon_release":20,"min_mon_release_name":"tentacle","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid","tentacle","nvmeof_beacon_diff"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:3300","nonce":0},{"type":"v1","addr":"192.168.123.107:6789","nonce":0}]},"addr":"192.168.123.107:6789/0","public_addr":"192.168.123.107:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1]} 2026-03-21T06:48:30.899 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 2 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cephadm 2026-03-21T06:48:25.278385+0000 mgr.x (mgr.14152) 36 : cephadm [INF] Deploying daemon mon.b on vm04 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cephadm 2026-03-21T06:48:25.278385+0000 mgr.x (mgr.14152) 36 : cephadm [INF] Deploying daemon mon.b on vm04 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:25.441939+0000 mon.a (mon.0) 159 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:25.441939+0000 mon.a (mon.0) 159 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:25.442139+0000 mon.a (mon.0) 160 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:25.442139+0000 mon.a (mon.0) 160 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:25.442253+0000 mon.a (mon.0) 161 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:25.442253+0000 mon.a (mon.0) 161 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:26.438924+0000 mon.a (mon.0) 162 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:26.438924+0000 mon.a (mon.0) 162 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:27.005379+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:27.005379+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:27.439344+0000 mon.a (mon.0) 164 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:27.439344+0000 mon.a (mon.0) 164 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:27.444989+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:27.444989+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:28.004918+0000 mon.a (mon.0) 165 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:28.004918+0000 mon.a (mon.0) 165 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:28.438505+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:28.438505+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:29.005037+0000 mon.a (mon.0) 167 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:29.005037+0000 mon.a (mon.0) 167 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:29.438371+0000 mon.a (mon.0) 168 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:29.438371+0000 mon.a (mon.0) 168 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:29.537781+0000 mgr.x (mgr.14152) 37 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:29.537781+0000 mgr.x (mgr.14152) 37 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.005157+0000 mon.a (mon.0) 169 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.005157+0000 mon.a (mon.0) 169 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.438749+0000 mon.a (mon.0) 170 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.438749+0000 mon.a (mon.0) 170 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.446776+0000 mon.a (mon.0) 171 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.446776+0000 mon.a (mon.0) 171 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450429+0000 mon.a (mon.0) 172 : cluster [DBG] monmap epoch 2 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450429+0000 mon.a (mon.0) 172 : cluster [DBG] monmap epoch 2 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450437+0000 mon.a (mon.0) 173 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450437+0000 mon.a (mon.0) 173 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450441+0000 mon.a (mon.0) 174 : cluster [DBG] last_changed 2026-03-21T06:48:25.438317+0000 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450441+0000 mon.a (mon.0) 174 : cluster [DBG] last_changed 2026-03-21T06:48:25.438317+0000 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450445+0000 mon.a (mon.0) 175 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450445+0000 mon.a (mon.0) 175 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450448+0000 mon.a (mon.0) 176 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450448+0000 mon.a (mon.0) 176 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450452+0000 mon.a (mon.0) 177 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450452+0000 mon.a (mon.0) 177 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450455+0000 mon.a (mon.0) 178 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450455+0000 mon.a (mon.0) 178 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450458+0000 mon.a (mon.0) 179 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450458+0000 mon.a (mon.0) 179 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450786+0000 mon.a (mon.0) 180 : cluster [DBG] fsmap 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450786+0000 mon.a (mon.0) 180 : cluster [DBG] fsmap 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450802+0000 mon.a (mon.0) 181 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450802+0000 mon.a (mon.0) 181 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450935+0000 mon.a (mon.0) 182 : cluster [DBG] mgrmap e13: x(active, since 20s) 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.450935+0000 mon.a (mon.0) 182 : cluster [DBG] mgrmap e13: x(active, since 20s) 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.451013+0000 mon.a (mon.0) 183 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: cluster 2026-03-21T06:48:30.451013+0000 mon.a (mon.0) 183 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.454382+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.454382+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.460349+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.460349+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.464229+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.464229+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.467207+0000 mon.a (mon.0) 187 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.467207+0000 mon.a (mon.0) 187 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.480753+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:30 vm07 bash[19945]: audit 2026-03-21T06:48:30.480753+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:31.896 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:31 vm02 bash[17927]: debug 2026-03-21T06:48:31.434+0000 7f1aef3d5640 -1 mgr.server handle_report got status from non-daemon mon.c 2026-03-21T06:48:31.951 INFO:tasks.cephadm:Waiting for 3 mons in monmap... 2026-03-21T06:48:31.951 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph mon dump -f json 2026-03-21T06:48:32.202 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:48:36.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:31.011120+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:36.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:31.011120+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:36.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:31.011366+0000 mon.a (mon.0) 192 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:36.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:31.011366+0000 mon.a (mon.0) 192 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:36.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:31.012441+0000 mon.a (mon.0) 193 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:31.012441+0000 mon.a (mon.0) 193 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:31.012739+0000 mon.a (mon.0) 194 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:31.012739+0000 mon.a (mon.0) 194 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:31.018025+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:31.018025+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:31.537970+0000 mgr.x (mgr.14152) 38 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:31.537970+0000 mgr.x (mgr.14152) 38 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:32.006032+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:32.006032+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:33.005957+0000 mon.a (mon.0) 196 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:33.005957+0000 mon.a (mon.0) 196 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:33.016553+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:33.016553+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:33.538178+0000 mgr.x (mgr.14152) 39 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:33.538178+0000 mgr.x (mgr.14152) 39 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:34.005693+0000 mon.a (mon.0) 197 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:34.005693+0000 mon.a (mon.0) 197 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:35.005480+0000 mon.a (mon.0) 198 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:35.005480+0000 mon.a (mon.0) 198 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:35.538343+0000 mgr.x (mgr.14152) 40 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:35.538343+0000 mgr.x (mgr.14152) 40 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.005862+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.005862+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.018538+0000 mon.a (mon.0) 200 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.018538+0000 mon.a (mon.0) 200 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022057+0000 mon.a (mon.0) 201 : cluster [DBG] monmap epoch 3 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022057+0000 mon.a (mon.0) 201 : cluster [DBG] monmap epoch 3 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022065+0000 mon.a (mon.0) 202 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022065+0000 mon.a (mon.0) 202 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022070+0000 mon.a (mon.0) 203 : cluster [DBG] last_changed 2026-03-21T06:48:31.006786+0000 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022070+0000 mon.a (mon.0) 203 : cluster [DBG] last_changed 2026-03-21T06:48:31.006786+0000 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022073+0000 mon.a (mon.0) 204 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022073+0000 mon.a (mon.0) 204 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022076+0000 mon.a (mon.0) 205 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022076+0000 mon.a (mon.0) 205 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022083+0000 mon.a (mon.0) 206 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022083+0000 mon.a (mon.0) 206 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022088+0000 mon.a (mon.0) 207 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022088+0000 mon.a (mon.0) 207 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022091+0000 mon.a (mon.0) 208 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022091+0000 mon.a (mon.0) 208 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022096+0000 mon.a (mon.0) 209 : cluster [DBG] 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022096+0000 mon.a (mon.0) 209 : cluster [DBG] 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022394+0000 mon.a (mon.0) 210 : cluster [DBG] fsmap 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022394+0000 mon.a (mon.0) 210 : cluster [DBG] fsmap 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022408+0000 mon.a (mon.0) 211 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022408+0000 mon.a (mon.0) 211 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022531+0000 mon.a (mon.0) 212 : cluster [DBG] mgrmap e13: x(active, since 26s) 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022531+0000 mon.a (mon.0) 212 : cluster [DBG] mgrmap e13: x(active, since 26s) 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022610+0000 mon.a (mon.0) 213 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: cluster 2026-03-21T06:48:36.022610+0000 mon.a (mon.0) 213 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.028079+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.028079+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.031967+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.031967+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.035597+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.035597+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.039658+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.039658+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.043057+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.043057+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.398 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.043762+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:36.398 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.043762+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:36.398 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.044293+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:36.398 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:36 vm02 bash[17657]: audit 2026-03-21T06:48:36.044293+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:36.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:31.011120+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:36.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:31.011120+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:36.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:31.011366+0000 mon.a (mon.0) 192 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:36.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:31.011366+0000 mon.a (mon.0) 192 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:36.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:31.012441+0000 mon.a (mon.0) 193 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:31.012441+0000 mon.a (mon.0) 193 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:31.012739+0000 mon.a (mon.0) 194 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:31.012739+0000 mon.a (mon.0) 194 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:31.018025+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:31.018025+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:31.537970+0000 mgr.x (mgr.14152) 38 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:31.537970+0000 mgr.x (mgr.14152) 38 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:32.006032+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:32.006032+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:33.005957+0000 mon.a (mon.0) 196 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:33.005957+0000 mon.a (mon.0) 196 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:33.016553+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:33.016553+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:33.538178+0000 mgr.x (mgr.14152) 39 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:33.538178+0000 mgr.x (mgr.14152) 39 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:34.005693+0000 mon.a (mon.0) 197 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:34.005693+0000 mon.a (mon.0) 197 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:35.005480+0000 mon.a (mon.0) 198 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:35.005480+0000 mon.a (mon.0) 198 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:35.538343+0000 mgr.x (mgr.14152) 40 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:35.538343+0000 mgr.x (mgr.14152) 40 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.005862+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.005862+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.018538+0000 mon.a (mon.0) 200 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.018538+0000 mon.a (mon.0) 200 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022057+0000 mon.a (mon.0) 201 : cluster [DBG] monmap epoch 3 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022057+0000 mon.a (mon.0) 201 : cluster [DBG] monmap epoch 3 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022065+0000 mon.a (mon.0) 202 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022065+0000 mon.a (mon.0) 202 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022070+0000 mon.a (mon.0) 203 : cluster [DBG] last_changed 2026-03-21T06:48:31.006786+0000 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022070+0000 mon.a (mon.0) 203 : cluster [DBG] last_changed 2026-03-21T06:48:31.006786+0000 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022073+0000 mon.a (mon.0) 204 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022073+0000 mon.a (mon.0) 204 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022076+0000 mon.a (mon.0) 205 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022076+0000 mon.a (mon.0) 205 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022083+0000 mon.a (mon.0) 206 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022083+0000 mon.a (mon.0) 206 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022088+0000 mon.a (mon.0) 207 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022088+0000 mon.a (mon.0) 207 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022091+0000 mon.a (mon.0) 208 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022091+0000 mon.a (mon.0) 208 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022096+0000 mon.a (mon.0) 209 : cluster [DBG] 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022096+0000 mon.a (mon.0) 209 : cluster [DBG] 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022394+0000 mon.a (mon.0) 210 : cluster [DBG] fsmap 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022394+0000 mon.a (mon.0) 210 : cluster [DBG] fsmap 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022408+0000 mon.a (mon.0) 211 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022408+0000 mon.a (mon.0) 211 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022531+0000 mon.a (mon.0) 212 : cluster [DBG] mgrmap e13: x(active, since 26s) 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022531+0000 mon.a (mon.0) 212 : cluster [DBG] mgrmap e13: x(active, since 26s) 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022610+0000 mon.a (mon.0) 213 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: cluster 2026-03-21T06:48:36.022610+0000 mon.a (mon.0) 213 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.028079+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.028079+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.031967+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.031967+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.035597+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.035597+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.039658+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.039658+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.043057+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.043057+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.043762+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.043762+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.044293+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:36.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:36 vm07 bash[19945]: audit 2026-03-21T06:48:36.044293+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:36.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cephadm 2026-03-21T06:48:25.278385+0000 mgr.x (mgr.14152) 36 : cephadm [INF] Deploying daemon mon.b on vm04 2026-03-21T06:48:36.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cephadm 2026-03-21T06:48:25.278385+0000 mgr.x (mgr.14152) 36 : cephadm [INF] Deploying daemon mon.b on vm04 2026-03-21T06:48:36.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:25.441939+0000 mon.a (mon.0) 159 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:36.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:25.441939+0000 mon.a (mon.0) 159 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:36.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:25.442139+0000 mon.a (mon.0) 160 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:25.442139+0000 mon.a (mon.0) 160 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:25.442253+0000 mon.a (mon.0) 161 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:36.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:25.442253+0000 mon.a (mon.0) 161 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:36.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:26.438924+0000 mon.a (mon.0) 162 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:26.438924+0000 mon.a (mon.0) 162 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:27.005379+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:27.005379+0000 mon.a (mon.0) 163 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:27.439344+0000 mon.a (mon.0) 164 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:27.439344+0000 mon.a (mon.0) 164 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:27.444989+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:27.444989+0000 mon.c (mon.1) 1 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:28.004918+0000 mon.a (mon.0) 165 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:28.004918+0000 mon.a (mon.0) 165 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:28.438505+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:28.438505+0000 mon.a (mon.0) 166 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:29.005037+0000 mon.a (mon.0) 167 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:29.005037+0000 mon.a (mon.0) 167 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:29.438371+0000 mon.a (mon.0) 168 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:29.438371+0000 mon.a (mon.0) 168 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:29.537781+0000 mgr.x (mgr.14152) 37 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:29.537781+0000 mgr.x (mgr.14152) 37 : cluster [DBG] pgmap v4: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.005157+0000 mon.a (mon.0) 169 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.005157+0000 mon.a (mon.0) 169 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.438749+0000 mon.a (mon.0) 170 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.438749+0000 mon.a (mon.0) 170 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.446776+0000 mon.a (mon.0) 171 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.446776+0000 mon.a (mon.0) 171 : cluster [INF] mon.a is new leader, mons a,c in quorum (ranks 0,1) 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450429+0000 mon.a (mon.0) 172 : cluster [DBG] monmap epoch 2 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450429+0000 mon.a (mon.0) 172 : cluster [DBG] monmap epoch 2 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450437+0000 mon.a (mon.0) 173 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450437+0000 mon.a (mon.0) 173 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450441+0000 mon.a (mon.0) 174 : cluster [DBG] last_changed 2026-03-21T06:48:25.438317+0000 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450441+0000 mon.a (mon.0) 174 : cluster [DBG] last_changed 2026-03-21T06:48:25.438317+0000 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450445+0000 mon.a (mon.0) 175 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450445+0000 mon.a (mon.0) 175 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450448+0000 mon.a (mon.0) 176 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450448+0000 mon.a (mon.0) 176 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450452+0000 mon.a (mon.0) 177 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450452+0000 mon.a (mon.0) 177 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450455+0000 mon.a (mon.0) 178 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450455+0000 mon.a (mon.0) 178 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450458+0000 mon.a (mon.0) 179 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450458+0000 mon.a (mon.0) 179 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450786+0000 mon.a (mon.0) 180 : cluster [DBG] fsmap 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450786+0000 mon.a (mon.0) 180 : cluster [DBG] fsmap 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450802+0000 mon.a (mon.0) 181 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450802+0000 mon.a (mon.0) 181 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450935+0000 mon.a (mon.0) 182 : cluster [DBG] mgrmap e13: x(active, since 20s) 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.450935+0000 mon.a (mon.0) 182 : cluster [DBG] mgrmap e13: x(active, since 20s) 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.451013+0000 mon.a (mon.0) 183 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:30.451013+0000 mon.a (mon.0) 183 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.454382+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.454382+0000 mon.a (mon.0) 184 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.460349+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.460349+0000 mon.a (mon.0) 185 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.464229+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.464229+0000 mon.a (mon.0) 186 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.467207+0000 mon.a (mon.0) 187 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.467207+0000 mon.a (mon.0) 187 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.480753+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:30.480753+0000 mon.a (mon.0) 188 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:36.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:31.011120+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:31.011120+0000 mon.a (mon.0) 191 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:31.011366+0000 mon.a (mon.0) 192 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:31.011366+0000 mon.a (mon.0) 192 : cluster [INF] mon.a calling monitor election 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:31.012441+0000 mon.a (mon.0) 193 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:31.012441+0000 mon.a (mon.0) 193 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:31.012739+0000 mon.a (mon.0) 194 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:31.012739+0000 mon.a (mon.0) 194 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:31.018025+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:31.018025+0000 mon.c (mon.1) 2 : cluster [INF] mon.c calling monitor election 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:31.537970+0000 mgr.x (mgr.14152) 38 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:31.537970+0000 mgr.x (mgr.14152) 38 : cluster [DBG] pgmap v5: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:32.006032+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:32.006032+0000 mon.a (mon.0) 195 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:33.005957+0000 mon.a (mon.0) 196 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:33.005957+0000 mon.a (mon.0) 196 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:33.016553+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:33.016553+0000 mon.b (mon.2) 1 : cluster [INF] mon.b calling monitor election 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:33.538178+0000 mgr.x (mgr.14152) 39 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:33.538178+0000 mgr.x (mgr.14152) 39 : cluster [DBG] pgmap v6: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:34.005693+0000 mon.a (mon.0) 197 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:34.005693+0000 mon.a (mon.0) 197 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:35.005480+0000 mon.a (mon.0) 198 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:35.005480+0000 mon.a (mon.0) 198 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:35.538343+0000 mgr.x (mgr.14152) 40 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:35.538343+0000 mgr.x (mgr.14152) 40 : cluster [DBG] pgmap v7: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.005862+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.005862+0000 mon.a (mon.0) 199 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.018538+0000 mon.a (mon.0) 200 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.018538+0000 mon.a (mon.0) 200 : cluster [INF] mon.a is new leader, mons a,c,b in quorum (ranks 0,1,2) 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022057+0000 mon.a (mon.0) 201 : cluster [DBG] monmap epoch 3 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022057+0000 mon.a (mon.0) 201 : cluster [DBG] monmap epoch 3 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022065+0000 mon.a (mon.0) 202 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022065+0000 mon.a (mon.0) 202 : cluster [DBG] fsid b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022070+0000 mon.a (mon.0) 203 : cluster [DBG] last_changed 2026-03-21T06:48:31.006786+0000 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022070+0000 mon.a (mon.0) 203 : cluster [DBG] last_changed 2026-03-21T06:48:31.006786+0000 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022073+0000 mon.a (mon.0) 204 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022073+0000 mon.a (mon.0) 204 : cluster [DBG] created 2026-03-21T06:47:46.690400+0000 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022076+0000 mon.a (mon.0) 205 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022076+0000 mon.a (mon.0) 205 : cluster [DBG] min_mon_release 20 (tentacle) 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022083+0000 mon.a (mon.0) 206 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022083+0000 mon.a (mon.0) 206 : cluster [DBG] election_strategy: 1 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022088+0000 mon.a (mon.0) 207 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022088+0000 mon.a (mon.0) 207 : cluster [DBG] 0: [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] mon.a 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022091+0000 mon.a (mon.0) 208 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022091+0000 mon.a (mon.0) 208 : cluster [DBG] 1: [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] mon.c 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022096+0000 mon.a (mon.0) 209 : cluster [DBG] 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022096+0000 mon.a (mon.0) 209 : cluster [DBG] 2: [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] mon.b 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022394+0000 mon.a (mon.0) 210 : cluster [DBG] fsmap 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022394+0000 mon.a (mon.0) 210 : cluster [DBG] fsmap 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022408+0000 mon.a (mon.0) 211 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022408+0000 mon.a (mon.0) 211 : cluster [DBG] osdmap e4: 0 total, 0 up, 0 in 2026-03-21T06:48:36.506 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022531+0000 mon.a (mon.0) 212 : cluster [DBG] mgrmap e13: x(active, since 26s) 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022531+0000 mon.a (mon.0) 212 : cluster [DBG] mgrmap e13: x(active, since 26s) 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022610+0000 mon.a (mon.0) 213 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: cluster 2026-03-21T06:48:36.022610+0000 mon.a (mon.0) 213 : cluster [INF] overall HEALTH_OK 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.028079+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.028079+0000 mon.a (mon.0) 214 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.031967+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.031967+0000 mon.a (mon.0) 215 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.035597+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.035597+0000 mon.a (mon.0) 216 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.039658+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.039658+0000 mon.a (mon.0) 217 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.043057+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.043057+0000 mon.a (mon.0) 218 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.043762+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.043762+0000 mon.a (mon.0) 219 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.044293+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:36.507 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:36 vm04 bash[20194]: audit 2026-03-21T06:48:36.044293+0000 mon.a (mon.0) 220 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:37.013 INFO:teuthology.orchestra.run.vm07.stderr:dumped monmap epoch 3 2026-03-21T06:48:37.013 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:48:37.013 INFO:teuthology.orchestra.run.vm07.stdout:{"epoch":3,"fsid":"b16ecafc-24f1-11f1-8ede-8330751617ee","modified":"2026-03-21T06:48:31.006786Z","created":"2026-03-21T06:47:46.690400Z","min_mon_release":20,"min_mon_release_name":"tentacle","election_strategy":1,"disallowed_leaders":"","stretch_mode":false,"tiebreaker_mon":"","removed_ranks":"","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus","octopus","pacific","elector-pinging","quincy","reef","squid","tentacle","nvmeof_beacon_diff"],"optional":[]},"mons":[{"rank":0,"name":"a","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:3300","nonce":0},{"type":"v1","addr":"192.168.123.102:6789","nonce":0}]},"addr":"192.168.123.102:6789/0","public_addr":"192.168.123.102:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":1,"name":"c","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:3300","nonce":0},{"type":"v1","addr":"192.168.123.107:6789","nonce":0}]},"addr":"192.168.123.107:6789/0","public_addr":"192.168.123.107:6789/0","priority":0,"weight":0,"crush_location":"{}"},{"rank":2,"name":"b","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:3300","nonce":0},{"type":"v1","addr":"192.168.123.104:6789","nonce":0}]},"addr":"192.168.123.104:6789/0","public_addr":"192.168.123.104:6789/0","priority":0,"weight":0,"crush_location":"{}"}],"quorum":[0,1,2]} 2026-03-21T06:48:37.074 INFO:tasks.cephadm:Generating final ceph.conf file... 2026-03-21T06:48:37.074 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph config generate-minimal-conf 2026-03-21T06:48:37.323 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:48:37.367 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.044880+0000 mgr.x (mgr.14152) 41 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:37.367 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.044880+0000 mgr.x (mgr.14152) 41 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:37.367 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.045170+0000 mgr.x (mgr.14152) 42 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:37.367 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.045170+0000 mgr.x (mgr.14152) 42 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:37.367 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.045484+0000 mgr.x (mgr.14152) 43 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:37.367 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.045484+0000 mgr.x (mgr.14152) 43 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:37.367 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.093720+0000 mgr.x (mgr.14152) 44 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.367 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.093720+0000 mgr.x (mgr.14152) 44 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.367 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.095483+0000 mgr.x (mgr.14152) 45 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.367 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.095483+0000 mgr.x (mgr.14152) 45 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.096382+0000 mgr.x (mgr.14152) 46 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.096382+0000 mgr.x (mgr.14152) 46 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.139998+0000 mon.a (mon.0) 221 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.139998+0000 mon.a (mon.0) 221 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.144159+0000 mon.a (mon.0) 222 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.144159+0000 mon.a (mon.0) 222 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.148010+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.148010+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.151627+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.151627+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.155875+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.155875+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.160197+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.160197+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.163819+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.163819+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.177339+0000 mgr.x (mgr.14152) 47 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.177339+0000 mgr.x (mgr.14152) 47 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.177597+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.177597+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.178126+0000 mon.a (mon.0) 229 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.178126+0000 mon.a (mon.0) 229 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.178530+0000 mon.a (mon.0) 230 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.178530+0000 mon.a (mon.0) 230 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.178929+0000 mgr.x (mgr.14152) 48 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.178929+0000 mgr.x (mgr.14152) 48 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.524577+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.524577+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.528921+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.528921+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.529515+0000 mgr.x (mgr.14152) 49 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.529515+0000 mgr.x (mgr.14152) 49 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.529699+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.529699+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.530215+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.530215+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.530638+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.530638+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.531077+0000 mgr.x (mgr.14152) 50 : cephadm [INF] Reconfiguring daemon mon.b on vm04 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.531077+0000 mgr.x (mgr.14152) 50 : cephadm [INF] Reconfiguring daemon mon.b on vm04 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.890589+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.890589+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.895358+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.895358+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.896392+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.896392+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.896962+0000 mon.a (mon.0) 239 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.896962+0000 mon.a (mon.0) 239 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.897415+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:36.897415+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:37.006029+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:37.006029+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:37.007510+0000 mon.a (mon.0) 242 : audit [DBG] from='client.? 192.168.123.107:0/1949217908' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:37.368 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:37 vm02 bash[17657]: audit 2026-03-21T06:48:37.007510+0000 mon.a (mon.0) 242 : audit [DBG] from='client.? 192.168.123.107:0/1949217908' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:37.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.044880+0000 mgr.x (mgr.14152) 41 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:37.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.044880+0000 mgr.x (mgr.14152) 41 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.045170+0000 mgr.x (mgr.14152) 42 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.045170+0000 mgr.x (mgr.14152) 42 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.045484+0000 mgr.x (mgr.14152) 43 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.045484+0000 mgr.x (mgr.14152) 43 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.093720+0000 mgr.x (mgr.14152) 44 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.093720+0000 mgr.x (mgr.14152) 44 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.095483+0000 mgr.x (mgr.14152) 45 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.095483+0000 mgr.x (mgr.14152) 45 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.096382+0000 mgr.x (mgr.14152) 46 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.096382+0000 mgr.x (mgr.14152) 46 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.139998+0000 mon.a (mon.0) 221 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.139998+0000 mon.a (mon.0) 221 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.144159+0000 mon.a (mon.0) 222 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.144159+0000 mon.a (mon.0) 222 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.148010+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.148010+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.151627+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.151627+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.155875+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.155875+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.160197+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.160197+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.163819+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.163819+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.177339+0000 mgr.x (mgr.14152) 47 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.177339+0000 mgr.x (mgr.14152) 47 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.177597+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.177597+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.178126+0000 mon.a (mon.0) 229 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.178126+0000 mon.a (mon.0) 229 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.178530+0000 mon.a (mon.0) 230 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.178530+0000 mon.a (mon.0) 230 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.178929+0000 mgr.x (mgr.14152) 48 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.178929+0000 mgr.x (mgr.14152) 48 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.524577+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.524577+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.528921+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.528921+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.529515+0000 mgr.x (mgr.14152) 49 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.529515+0000 mgr.x (mgr.14152) 49 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.529699+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.529699+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.530215+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.530215+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.530638+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.530638+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.531077+0000 mgr.x (mgr.14152) 50 : cephadm [INF] Reconfiguring daemon mon.b on vm04 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.531077+0000 mgr.x (mgr.14152) 50 : cephadm [INF] Reconfiguring daemon mon.b on vm04 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.890589+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.890589+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.895358+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.895358+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.896392+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.896392+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.896962+0000 mon.a (mon.0) 239 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.896962+0000 mon.a (mon.0) 239 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.897415+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:36.897415+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:37.006029+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:37.006029+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:37.007510+0000 mon.a (mon.0) 242 : audit [DBG] from='client.? 192.168.123.107:0/1949217908' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:37.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:37 vm07 bash[19945]: audit 2026-03-21T06:48:37.007510+0000 mon.a (mon.0) 242 : audit [DBG] from='client.? 192.168.123.107:0/1949217908' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.044880+0000 mgr.x (mgr.14152) 41 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.044880+0000 mgr.x (mgr.14152) 41 : cephadm [INF] Updating vm02:/etc/ceph/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.045170+0000 mgr.x (mgr.14152) 42 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.045170+0000 mgr.x (mgr.14152) 42 : cephadm [INF] Updating vm04:/etc/ceph/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.045484+0000 mgr.x (mgr.14152) 43 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.045484+0000 mgr.x (mgr.14152) 43 : cephadm [INF] Updating vm07:/etc/ceph/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.093720+0000 mgr.x (mgr.14152) 44 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.093720+0000 mgr.x (mgr.14152) 44 : cephadm [INF] Updating vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.095483+0000 mgr.x (mgr.14152) 45 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.095483+0000 mgr.x (mgr.14152) 45 : cephadm [INF] Updating vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.096382+0000 mgr.x (mgr.14152) 46 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.096382+0000 mgr.x (mgr.14152) 46 : cephadm [INF] Updating vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/config/ceph.conf 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.139998+0000 mon.a (mon.0) 221 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.139998+0000 mon.a (mon.0) 221 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.144159+0000 mon.a (mon.0) 222 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.144159+0000 mon.a (mon.0) 222 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.148010+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.148010+0000 mon.a (mon.0) 223 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.151627+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.151627+0000 mon.a (mon.0) 224 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.155875+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.155875+0000 mon.a (mon.0) 225 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.160197+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.160197+0000 mon.a (mon.0) 226 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.163819+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.163819+0000 mon.a (mon.0) 227 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.177339+0000 mgr.x (mgr.14152) 47 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-21T06:48:37.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.177339+0000 mgr.x (mgr.14152) 47 : cephadm [INF] Reconfiguring mon.a (unknown last config time)... 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.177597+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.177597+0000 mon.a (mon.0) 228 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.178126+0000 mon.a (mon.0) 229 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.178126+0000 mon.a (mon.0) 229 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.178530+0000 mon.a (mon.0) 230 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.178530+0000 mon.a (mon.0) 230 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.178929+0000 mgr.x (mgr.14152) 48 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.178929+0000 mgr.x (mgr.14152) 48 : cephadm [INF] Reconfiguring daemon mon.a on vm02 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.524577+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.524577+0000 mon.a (mon.0) 231 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.528921+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.528921+0000 mon.a (mon.0) 232 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.529515+0000 mgr.x (mgr.14152) 49 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.529515+0000 mgr.x (mgr.14152) 49 : cephadm [INF] Reconfiguring mon.b (monmap changed)... 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.529699+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.529699+0000 mon.a (mon.0) 233 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.530215+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.530215+0000 mon.a (mon.0) 234 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.530638+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.530638+0000 mon.a (mon.0) 235 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.531077+0000 mgr.x (mgr.14152) 50 : cephadm [INF] Reconfiguring daemon mon.b on vm04 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.531077+0000 mgr.x (mgr.14152) 50 : cephadm [INF] Reconfiguring daemon mon.b on vm04 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.890589+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.890589+0000 mon.a (mon.0) 236 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.895358+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.895358+0000 mon.a (mon.0) 237 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.896392+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.896392+0000 mon.a (mon.0) 238 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "mon."} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.896962+0000 mon.a (mon.0) 239 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.896962+0000 mon.a (mon.0) 239 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config get", "who": "mon", "key": "public_network"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.897415+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:36.897415+0000 mon.a (mon.0) 240 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:37.006029+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:37.006029+0000 mon.a (mon.0) 241 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:37.007510+0000 mon.a (mon.0) 242 : audit [DBG] from='client.? 192.168.123.107:0/1949217908' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:37.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:37 vm04 bash[20194]: audit 2026-03-21T06:48:37.007510+0000 mon.a (mon.0) 242 : audit [DBG] from='client.? 192.168.123.107:0/1949217908' entity='client.admin' cmd={"prefix": "mon dump", "format": "json"} : dispatch 2026-03-21T06:48:37.652 INFO:teuthology.orchestra.run.vm02.stdout:# minimal ceph.conf for b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:37.653 INFO:teuthology.orchestra.run.vm02.stdout:[global] 2026-03-21T06:48:37.653 INFO:teuthology.orchestra.run.vm02.stdout: fsid = b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T06:48:37.653 INFO:teuthology.orchestra.run.vm02.stdout: mon_host = [v2:192.168.123.102:3300/0,v1:192.168.123.102:6789/0] [v2:192.168.123.104:3300/0,v1:192.168.123.104:6789/0] [v2:192.168.123.107:3300/0,v1:192.168.123.107:6789/0] 2026-03-21T06:48:37.705 INFO:tasks.cephadm:Distributing (final) config and client.admin keyring... 2026-03-21T06:48:37.705 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:48:37.705 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.conf 2026-03-21T06:48:37.711 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:48:37.712 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:37.765 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:48:37.765 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.conf 2026-03-21T06:48:37.773 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:48:37.773 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:37.823 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:48:37.823 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/ceph/ceph.conf 2026-03-21T06:48:37.831 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:48:37.831 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/ceph/ceph.client.admin.keyring 2026-03-21T06:48:37.881 INFO:tasks.cephadm:Adding mgr.x on vm02 2026-03-21T06:48:37.881 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch apply mgr '1;vm02=x' 2026-03-21T06:48:38.150 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:48:38.473 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled mgr update... 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.896044+0000 mgr.x (mgr.14152) 51 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.896044+0000 mgr.x (mgr.14152) 51 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.897974+0000 mgr.x (mgr.14152) 52 : cephadm [INF] Reconfiguring daemon mon.c on vm07 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: cephadm 2026-03-21T06:48:36.897974+0000 mgr.x (mgr.14152) 52 : cephadm [INF] Reconfiguring daemon mon.c on vm07 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.266069+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.266069+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.270087+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.270087+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.270989+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.270989+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.271922+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.271922+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.272379+0000 mon.a (mon.0) 247 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.272379+0000 mon.a (mon.0) 247 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.275791+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.275791+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: cluster 2026-03-21T06:48:37.538534+0000 mgr.x (mgr.14152) 53 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: cluster 2026-03-21T06:48:37.538534+0000 mgr.x (mgr.14152) 53 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.652546+0000 mon.c (mon.1) 3 : audit [DBG] from='client.? 192.168.123.102:0/1794379708' entity='client.admin' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.490 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:38 vm07 bash[19945]: audit 2026-03-21T06:48:37.652546+0000 mon.c (mon.1) 3 : audit [DBG] from='client.? 192.168.123.102:0/1794379708' entity='client.admin' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.540 DEBUG:tasks.cephadm:set 0 configs 2026-03-21T06:48:38.541 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph config dump 2026-03-21T06:48:38.546 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.896044+0000 mgr.x (mgr.14152) 51 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-21T06:48:38.546 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.896044+0000 mgr.x (mgr.14152) 51 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-21T06:48:38.546 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.897974+0000 mgr.x (mgr.14152) 52 : cephadm [INF] Reconfiguring daemon mon.c on vm07 2026-03-21T06:48:38.546 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: cephadm 2026-03-21T06:48:36.897974+0000 mgr.x (mgr.14152) 52 : cephadm [INF] Reconfiguring daemon mon.c on vm07 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.266069+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.266069+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.270087+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.270087+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.270989+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.270989+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.271922+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.271922+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.272379+0000 mon.a (mon.0) 247 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.272379+0000 mon.a (mon.0) 247 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.275791+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.275791+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: cluster 2026-03-21T06:48:37.538534+0000 mgr.x (mgr.14152) 53 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: cluster 2026-03-21T06:48:37.538534+0000 mgr.x (mgr.14152) 53 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.652546+0000 mon.c (mon.1) 3 : audit [DBG] from='client.? 192.168.123.102:0/1794379708' entity='client.admin' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.547 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:38 vm02 bash[17657]: audit 2026-03-21T06:48:37.652546+0000 mon.c (mon.1) 3 : audit [DBG] from='client.? 192.168.123.102:0/1794379708' entity='client.admin' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.896044+0000 mgr.x (mgr.14152) 51 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.896044+0000 mgr.x (mgr.14152) 51 : cephadm [INF] Reconfiguring mon.c (monmap changed)... 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.897974+0000 mgr.x (mgr.14152) 52 : cephadm [INF] Reconfiguring daemon mon.c on vm07 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: cephadm 2026-03-21T06:48:36.897974+0000 mgr.x (mgr.14152) 52 : cephadm [INF] Reconfiguring daemon mon.c on vm07 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.266069+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.266069+0000 mon.a (mon.0) 243 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.270087+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.270087+0000 mon.a (mon.0) 244 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.270989+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.270989+0000 mon.a (mon.0) 245 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.271922+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.271922+0000 mon.a (mon.0) 246 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.272379+0000 mon.a (mon.0) 247 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.272379+0000 mon.a (mon.0) 247 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.275791+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.275791+0000 mon.a (mon.0) 248 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: cluster 2026-03-21T06:48:37.538534+0000 mgr.x (mgr.14152) 53 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: cluster 2026-03-21T06:48:37.538534+0000 mgr.x (mgr.14152) 53 : cluster [DBG] pgmap v8: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.652546+0000 mon.c (mon.1) 3 : audit [DBG] from='client.? 192.168.123.102:0/1794379708' entity='client.admin' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:38 vm04 bash[20194]: audit 2026-03-21T06:48:37.652546+0000 mon.c (mon.1) 3 : audit [DBG] from='client.? 192.168.123.102:0/1794379708' entity='client.admin' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:38.799 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:WHO MASK LEVEL OPTION VALUE RO 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global dev auth_debug true 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global basic container_image quay.ceph.io/ceph-ci/ceph@sha256:5fe4cfc092321d85e013fc971c0a15337be7ad4a3c8cc47ec11af07596687bd6 * 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global dev debug_asserts_on_shutdown true 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global basic log_to_file true 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global basic log_to_journald false 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global basic log_to_stderr false 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced mon_allow_pool_delete true 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced mon_clock_drift_allowed 1.000000 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced mon_cluster_log_to_file true 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced mon_max_pg_per_osd 10000 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced mon_pg_warn_max_object_skew 0.000000 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced mon_warn_on_crush_straw_calc_version_zero false 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced mon_warn_on_legacy_crush_tunables false 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced mon_warn_on_osd_down_out_interval_zero false 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global dev mon_warn_on_pool_no_app false 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global dev mon_warn_on_pool_pg_num_not_power_of_two false 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced mon_warn_on_too_few_osds false 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global dev ms_die_on_bug true 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global dev ms_die_on_old_message true 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced osd_pool_default_erasure_code_profile plugin=isa technique=reed_sol_van k=2 m=1 crush-failure-domain=osd 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced osd_pool_default_pg_autoscale_mode off 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:global advanced public_network 192.168.123.0/24,192.168.123.1/32 * 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced auth_allow_insecure_global_id_reclaim false 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced auth_mon_ticket_ttl 660.000000 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced auth_service_ticket_ttl 240.000000 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced debug_mon 20/20 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced debug_ms 1/1 2026-03-21T06:48:39.132 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced debug_paxos 20/20 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced mon_data_avail_warn 5 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced mon_mgr_mkfs_grace 240 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mon dev mon_osd_prime_pg_temp true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced mon_osd_reporter_subtree_level osd 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced mon_reweight_min_bytes_per_osd 10 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced mon_reweight_min_pgs_per_osd 4 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced mon_warn_on_insecure_global_id_reclaim false 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mon advanced mon_warn_on_insecure_global_id_reclaim_allowed false 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mgr advanced debug_mgr 20/20 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mgr advanced debug_ms 1/1 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mgr advanced mgr/cephadm/allow_ptrace true * 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mgr advanced mgr/cephadm/container_init True * 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mgr advanced mgr/cephadm/migration_current 7 * 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mgr advanced mgr/dashboard/ssl_server_port 8443 * 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mgr advanced mgr/orchestrator/orchestrator cephadm 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mgr advanced mon_reweight_min_bytes_per_osd 10 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:mgr advanced mon_reweight_min_pgs_per_osd 4 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd dev bdev_debug_aio true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced debug_ms 1/1 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced debug_osd 20/20 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd dev osd_debug_misdirected_ops true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd dev osd_debug_op_order true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd dev osd_debug_pg_log_writeout true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd dev osd_debug_shutdown true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd dev osd_debug_verify_cached_snaps true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd dev osd_debug_verify_missing_on_start true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd dev osd_debug_verify_stray_on_activate true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced osd_deep_scrub_update_digest_min_age 30 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd basic osd_mclock_iops_capacity_threshold_hdd 49000.000000 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced osd_mclock_profile high_recovery_ops 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd dev osd_mclock_skip_benchmark true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced osd_memory_target_autotune true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced osd_op_queue debug_random * 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced osd_op_queue_cut_off debug_random * 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced osd_recover_clone_overlap true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced osd_recovery_max_chunk 1048576 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced osd_scrub_load_threshold 5.000000 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced osd_scrub_max_interval 600.000000 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:osd advanced osd_shutdown_pgref_assert true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:client.rgw advanced rgw_cache_enabled true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:client.rgw advanced rgw_enable_ops_log true 2026-03-21T06:48:39.133 INFO:teuthology.orchestra.run.vm02.stdout:client.rgw advanced rgw_enable_usage_log true 2026-03-21T06:48:39.208 INFO:tasks.cephadm:Deploying OSDs... 2026-03-21T06:48:39.208 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:48:39.208 DEBUG:teuthology.orchestra.run.vm02:> dd if=/scratch_devs of=/dev/stdout 2026-03-21T06:48:39.255 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-03-21T06:48:39.255 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_1 2026-03-21T06:48:39.299 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-03-21T06:48:39.299 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:39.299 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 782 Links: 1 2026-03-21T06:48:39.299 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:39.299 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-21 06:46:14.114285000 +0000 2026-03-21T06:48:39.300 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-21 06:46:13.990285000 +0000 2026-03-21T06:48:39.300 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-21 06:46:13.990285000 +0000 2026-03-21T06:48:39.300 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-21T06:48:39.300 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-03-21T06:48:39.351 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-21T06:48:39.351 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-21T06:48:39.351 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000200184 s, 2.6 MB/s 2026-03-21T06:48:39.352 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-03-21T06:48:39.400 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_2 2026-03-21T06:48:39.447 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-03-21T06:48:39.447 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:39.447 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 814 Links: 1 2026-03-21T06:48:39.447 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:39.447 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-21 06:46:14.390285000 +0000 2026-03-21T06:48:39.447 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-21 06:46:14.274285000 +0000 2026-03-21T06:48:39.447 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-21 06:46:14.274285000 +0000 2026-03-21T06:48:39.447 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-21T06:48:39.447 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-03-21T06:48:39.477 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.463593+0000 mgr.x (mgr.14152) 54 : audit [DBG] from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm02=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:39.477 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.463593+0000 mgr.x (mgr.14152) 54 : audit [DBG] from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm02=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:39.477 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: cephadm 2026-03-21T06:48:38.464431+0000 mgr.x (mgr.14152) 55 : cephadm [INF] Saving service mgr spec with placement vm02=x;count:1 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: cephadm 2026-03-21T06:48:38.464431+0000 mgr.x (mgr.14152) 55 : cephadm [INF] Saving service mgr spec with placement vm02=x;count:1 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.467750+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.467750+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.468371+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.468371+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.469495+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.469495+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.469958+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.469958+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.473435+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.473435+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.476628+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.476628+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: cephadm 2026-03-21T06:48:38.491414+0000 mgr.x (mgr.14152) 56 : cephadm [INF] Reconfiguring mgr.x (unknown last config time)... 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: cephadm 2026-03-21T06:48:38.491414+0000 mgr.x (mgr.14152) 56 : cephadm [INF] Reconfiguring mgr.x (unknown last config time)... 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.491607+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.491607+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.492186+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr services"} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.492186+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr services"} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.492669+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.492669+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: cephadm 2026-03-21T06:48:38.493139+0000 mgr.x (mgr.14152) 57 : cephadm [INF] Reconfiguring daemon mgr.x on vm02 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: cephadm 2026-03-21T06:48:38.493139+0000 mgr.x (mgr.14152) 57 : cephadm [INF] Reconfiguring daemon mgr.x on vm02 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.926556+0000 mon.a (mon.0) 258 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.926556+0000 mon.a (mon.0) 258 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.934534+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:38.934534+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:39.126386+0000 mon.a (mon.0) 260 : audit [DBG] from='client.? 192.168.123.102:0/2840351519' entity='client.admin' cmd={"prefix": "config dump"} : dispatch 2026-03-21T06:48:39.478 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:39 vm02 bash[17657]: audit 2026-03-21T06:48:39.126386+0000 mon.a (mon.0) 260 : audit [DBG] from='client.? 192.168.123.102:0/2840351519' entity='client.admin' cmd={"prefix": "config dump"} : dispatch 2026-03-21T06:48:39.487 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-21T06:48:39.487 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-21T06:48:39.487 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000137667 s, 3.7 MB/s 2026-03-21T06:48:39.487 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-03-21T06:48:39.532 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_3 2026-03-21T06:48:39.579 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-03-21T06:48:39.579 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:39.579 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 844 Links: 1 2026-03-21T06:48:39.579 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:39.579 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-21 06:46:14.674285000 +0000 2026-03-21T06:48:39.579 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-21 06:46:14.550285000 +0000 2026-03-21T06:48:39.580 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-21 06:46:14.550285000 +0000 2026-03-21T06:48:39.580 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-21T06:48:39.580 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-03-21T06:48:39.627 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-21T06:48:39.627 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-21T06:48:39.627 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000190937 s, 2.7 MB/s 2026-03-21T06:48:39.628 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-03-21T06:48:39.677 DEBUG:teuthology.orchestra.run.vm02:> stat /dev/vg_nvme/lv_4 2026-03-21T06:48:39.723 INFO:teuthology.orchestra.run.vm02.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-03-21T06:48:39.723 INFO:teuthology.orchestra.run.vm02.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:39.723 INFO:teuthology.orchestra.run.vm02.stdout:Device: 5h/5d Inode: 874 Links: 1 2026-03-21T06:48:39.723 INFO:teuthology.orchestra.run.vm02.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:39.724 INFO:teuthology.orchestra.run.vm02.stdout:Access: 2026-03-21 06:46:18.162285000 +0000 2026-03-21T06:48:39.724 INFO:teuthology.orchestra.run.vm02.stdout:Modify: 2026-03-21 06:46:14.822285000 +0000 2026-03-21T06:48:39.724 INFO:teuthology.orchestra.run.vm02.stdout:Change: 2026-03-21 06:46:14.822285000 +0000 2026-03-21T06:48:39.724 INFO:teuthology.orchestra.run.vm02.stdout: Birth: - 2026-03-21T06:48:39.724 DEBUG:teuthology.orchestra.run.vm02:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.463593+0000 mgr.x (mgr.14152) 54 : audit [DBG] from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm02=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.463593+0000 mgr.x (mgr.14152) 54 : audit [DBG] from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm02=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: cephadm 2026-03-21T06:48:38.464431+0000 mgr.x (mgr.14152) 55 : cephadm [INF] Saving service mgr spec with placement vm02=x;count:1 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: cephadm 2026-03-21T06:48:38.464431+0000 mgr.x (mgr.14152) 55 : cephadm [INF] Saving service mgr spec with placement vm02=x;count:1 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.467750+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.467750+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.468371+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.468371+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.469495+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.469495+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.469958+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.469958+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.473435+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.473435+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.476628+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.476628+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: cephadm 2026-03-21T06:48:38.491414+0000 mgr.x (mgr.14152) 56 : cephadm [INF] Reconfiguring mgr.x (unknown last config time)... 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: cephadm 2026-03-21T06:48:38.491414+0000 mgr.x (mgr.14152) 56 : cephadm [INF] Reconfiguring mgr.x (unknown last config time)... 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.491607+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.491607+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.492186+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr services"} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.492186+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr services"} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.492669+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.492669+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: cephadm 2026-03-21T06:48:38.493139+0000 mgr.x (mgr.14152) 57 : cephadm [INF] Reconfiguring daemon mgr.x on vm02 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: cephadm 2026-03-21T06:48:38.493139+0000 mgr.x (mgr.14152) 57 : cephadm [INF] Reconfiguring daemon mgr.x on vm02 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.926556+0000 mon.a (mon.0) 258 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.926556+0000 mon.a (mon.0) 258 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.934534+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:38.934534+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:39.126386+0000 mon.a (mon.0) 260 : audit [DBG] from='client.? 192.168.123.102:0/2840351519' entity='client.admin' cmd={"prefix": "config dump"} : dispatch 2026-03-21T06:48:39.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:39 vm04 bash[20194]: audit 2026-03-21T06:48:39.126386+0000 mon.a (mon.0) 260 : audit [DBG] from='client.? 192.168.123.102:0/2840351519' entity='client.admin' cmd={"prefix": "config dump"} : dispatch 2026-03-21T06:48:39.777 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records in 2026-03-21T06:48:39.777 INFO:teuthology.orchestra.run.vm02.stderr:1+0 records out 2026-03-21T06:48:39.777 INFO:teuthology.orchestra.run.vm02.stderr:512 bytes copied, 0.000166552 s, 3.1 MB/s 2026-03-21T06:48:39.777 DEBUG:teuthology.orchestra.run.vm02:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-03-21T06:48:39.824 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:48:39.824 DEBUG:teuthology.orchestra.run.vm04:> dd if=/scratch_devs of=/dev/stdout 2026-03-21T06:48:39.827 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-03-21T06:48:39.827 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vg_nvme/lv_1 2026-03-21T06:48:39.871 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-03-21T06:48:39.871 INFO:teuthology.orchestra.run.vm04.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:39.871 INFO:teuthology.orchestra.run.vm04.stdout:Device: 5h/5d Inode: 778 Links: 1 2026-03-21T06:48:39.871 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:39.871 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-21 06:46:00.038604000 +0000 2026-03-21T06:48:39.871 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-21 06:45:59.902604000 +0000 2026-03-21T06:48:39.871 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-21 06:45:59.902604000 +0000 2026-03-21T06:48:39.871 INFO:teuthology.orchestra.run.vm04.stdout: Birth: - 2026-03-21T06:48:39.871 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-03-21T06:48:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.463593+0000 mgr.x (mgr.14152) 54 : audit [DBG] from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm02=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.463593+0000 mgr.x (mgr.14152) 54 : audit [DBG] from='client.14214 -' entity='client.admin' cmd=[{"prefix": "orch apply", "service_type": "mgr", "placement": "1;vm02=x", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: cephadm 2026-03-21T06:48:38.464431+0000 mgr.x (mgr.14152) 55 : cephadm [INF] Saving service mgr spec with placement vm02=x;count:1 2026-03-21T06:48:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: cephadm 2026-03-21T06:48:38.464431+0000 mgr.x (mgr.14152) 55 : cephadm [INF] Saving service mgr spec with placement vm02=x;count:1 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.467750+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.467750+0000 mon.a (mon.0) 249 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.468371+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.468371+0000 mon.a (mon.0) 250 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.469495+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.469495+0000 mon.a (mon.0) 251 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.469958+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.469958+0000 mon.a (mon.0) 252 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.473435+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.473435+0000 mon.a (mon.0) 253 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.476628+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.476628+0000 mon.a (mon.0) 254 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: cephadm 2026-03-21T06:48:38.491414+0000 mgr.x (mgr.14152) 56 : cephadm [INF] Reconfiguring mgr.x (unknown last config time)... 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: cephadm 2026-03-21T06:48:38.491414+0000 mgr.x (mgr.14152) 56 : cephadm [INF] Reconfiguring mgr.x (unknown last config time)... 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.491607+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.491607+0000 mon.a (mon.0) 255 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "mgr.x", "caps": ["mon", "profile mgr", "osd", "allow *", "mds", "allow *"]} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.492186+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr services"} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.492186+0000 mon.a (mon.0) 256 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mgr services"} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.492669+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.492669+0000 mon.a (mon.0) 257 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: cephadm 2026-03-21T06:48:38.493139+0000 mgr.x (mgr.14152) 57 : cephadm [INF] Reconfiguring daemon mgr.x on vm02 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: cephadm 2026-03-21T06:48:38.493139+0000 mgr.x (mgr.14152) 57 : cephadm [INF] Reconfiguring daemon mgr.x on vm02 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.926556+0000 mon.a (mon.0) 258 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.926556+0000 mon.a (mon.0) 258 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.934534+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:38.934534+0000 mon.a (mon.0) 259 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:39.126386+0000 mon.a (mon.0) 260 : audit [DBG] from='client.? 192.168.123.102:0/2840351519' entity='client.admin' cmd={"prefix": "config dump"} : dispatch 2026-03-21T06:48:39.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:39 vm07 bash[19945]: audit 2026-03-21T06:48:39.126386+0000 mon.a (mon.0) 260 : audit [DBG] from='client.? 192.168.123.102:0/2840351519' entity='client.admin' cmd={"prefix": "config dump"} : dispatch 2026-03-21T06:48:39.919 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-21T06:48:39.919 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-21T06:48:39.919 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000154079 s, 3.3 MB/s 2026-03-21T06:48:39.920 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-03-21T06:48:39.964 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vg_nvme/lv_2 2026-03-21T06:48:40.011 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-03-21T06:48:40.011 INFO:teuthology.orchestra.run.vm04.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:40.011 INFO:teuthology.orchestra.run.vm04.stdout:Device: 5h/5d Inode: 809 Links: 1 2026-03-21T06:48:40.011 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:40.012 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-21 06:46:00.342604000 +0000 2026-03-21T06:48:40.012 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-21 06:46:00.198604000 +0000 2026-03-21T06:48:40.012 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-21 06:46:00.198604000 +0000 2026-03-21T06:48:40.012 INFO:teuthology.orchestra.run.vm04.stdout: Birth: - 2026-03-21T06:48:40.012 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-03-21T06:48:40.060 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-21T06:48:40.060 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-21T06:48:40.060 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000169176 s, 3.0 MB/s 2026-03-21T06:48:40.061 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-03-21T06:48:40.109 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vg_nvme/lv_3 2026-03-21T06:48:40.156 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-03-21T06:48:40.156 INFO:teuthology.orchestra.run.vm04.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:40.156 INFO:teuthology.orchestra.run.vm04.stdout:Device: 5h/5d Inode: 840 Links: 1 2026-03-21T06:48:40.156 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:40.156 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-21 06:46:00.662604000 +0000 2026-03-21T06:48:40.156 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-21 06:46:00.518604000 +0000 2026-03-21T06:48:40.156 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-21 06:46:00.518604000 +0000 2026-03-21T06:48:40.156 INFO:teuthology.orchestra.run.vm04.stdout: Birth: - 2026-03-21T06:48:40.156 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-03-21T06:48:40.203 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-21T06:48:40.203 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-21T06:48:40.203 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000166241 s, 3.1 MB/s 2026-03-21T06:48:40.204 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-03-21T06:48:40.248 DEBUG:teuthology.orchestra.run.vm04:> stat /dev/vg_nvme/lv_4 2026-03-21T06:48:40.291 INFO:teuthology.orchestra.run.vm04.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-03-21T06:48:40.291 INFO:teuthology.orchestra.run.vm04.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:40.291 INFO:teuthology.orchestra.run.vm04.stdout:Device: 5h/5d Inode: 871 Links: 1 2026-03-21T06:48:40.291 INFO:teuthology.orchestra.run.vm04.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:40.291 INFO:teuthology.orchestra.run.vm04.stdout:Access: 2026-03-21 06:46:18.182604000 +0000 2026-03-21T06:48:40.291 INFO:teuthology.orchestra.run.vm04.stdout:Modify: 2026-03-21 06:46:00.818604000 +0000 2026-03-21T06:48:40.291 INFO:teuthology.orchestra.run.vm04.stdout:Change: 2026-03-21 06:46:00.818604000 +0000 2026-03-21T06:48:40.291 INFO:teuthology.orchestra.run.vm04.stdout: Birth: - 2026-03-21T06:48:40.291 DEBUG:teuthology.orchestra.run.vm04:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-03-21T06:48:40.339 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records in 2026-03-21T06:48:40.339 INFO:teuthology.orchestra.run.vm04.stderr:1+0 records out 2026-03-21T06:48:40.339 INFO:teuthology.orchestra.run.vm04.stderr:512 bytes copied, 0.000157023 s, 3.3 MB/s 2026-03-21T06:48:40.340 DEBUG:teuthology.orchestra.run.vm04:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-03-21T06:48:40.384 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:48:40.384 DEBUG:teuthology.orchestra.run.vm07:> dd if=/scratch_devs of=/dev/stdout 2026-03-21T06:48:40.387 DEBUG:teuthology.misc:devs=['/dev/vg_nvme/lv_1', '/dev/vg_nvme/lv_2', '/dev/vg_nvme/lv_3', '/dev/vg_nvme/lv_4'] 2026-03-21T06:48:40.387 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vg_nvme/lv_1 2026-03-21T06:48:40.433 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vg_nvme/lv_1 -> ../dm-0 2026-03-21T06:48:40.433 INFO:teuthology.orchestra.run.vm07.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:40.433 INFO:teuthology.orchestra.run.vm07.stdout:Device: 5h/5d Inode: 793 Links: 1 2026-03-21T06:48:40.433 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:40.433 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-21 06:45:50.781309000 +0000 2026-03-21T06:48:40.433 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-21 06:45:50.641309000 +0000 2026-03-21T06:48:40.433 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-21 06:45:50.641309000 +0000 2026-03-21T06:48:40.433 INFO:teuthology.orchestra.run.vm07.stdout: Birth: - 2026-03-21T06:48:40.433 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vg_nvme/lv_1 of=/dev/null count=1 2026-03-21T06:48:40.480 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-21T06:48:40.480 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-21T06:48:40.480 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.000114374 s, 4.5 MB/s 2026-03-21T06:48:40.481 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_1 2026-03-21T06:48:40.526 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vg_nvme/lv_2 2026-03-21T06:48:40.569 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vg_nvme/lv_2 -> ../dm-1 2026-03-21T06:48:40.569 INFO:teuthology.orchestra.run.vm07.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:40.569 INFO:teuthology.orchestra.run.vm07.stdout:Device: 5h/5d Inode: 822 Links: 1 2026-03-21T06:48:40.569 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:40.569 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-21 06:45:51.089309000 +0000 2026-03-21T06:48:40.569 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-21 06:45:50.953309000 +0000 2026-03-21T06:48:40.569 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-21 06:45:50.953309000 +0000 2026-03-21T06:48:40.569 INFO:teuthology.orchestra.run.vm07.stdout: Birth: - 2026-03-21T06:48:40.569 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vg_nvme/lv_2 of=/dev/null count=1 2026-03-21T06:48:40.617 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-21T06:48:40.617 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-21T06:48:40.617 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.000146143 s, 3.5 MB/s 2026-03-21T06:48:40.618 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_2 2026-03-21T06:48:40.662 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vg_nvme/lv_3 2026-03-21T06:48:40.709 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vg_nvme/lv_3 -> ../dm-2 2026-03-21T06:48:40.709 INFO:teuthology.orchestra.run.vm07.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:40.709 INFO:teuthology.orchestra.run.vm07.stdout:Device: 5h/5d Inode: 854 Links: 1 2026-03-21T06:48:40.709 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:40.709 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-21 06:45:51.249309000 +0000 2026-03-21T06:48:40.709 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-21 06:45:51.245309000 +0000 2026-03-21T06:48:40.709 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-21 06:45:51.245309000 +0000 2026-03-21T06:48:40.709 INFO:teuthology.orchestra.run.vm07.stdout: Birth: - 2026-03-21T06:48:40.709 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vg_nvme/lv_3 of=/dev/null count=1 2026-03-21T06:48:40.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:40 vm04 bash[20194]: cluster 2026-03-21T06:48:39.538735+0000 mgr.x (mgr.14152) 58 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:40.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:40 vm04 bash[20194]: cluster 2026-03-21T06:48:39.538735+0000 mgr.x (mgr.14152) 58 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:40.755 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:40 vm07 bash[19945]: cluster 2026-03-21T06:48:39.538735+0000 mgr.x (mgr.14152) 58 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:40.755 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:40 vm07 bash[19945]: cluster 2026-03-21T06:48:39.538735+0000 mgr.x (mgr.14152) 58 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:40.757 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-21T06:48:40.757 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-21T06:48:40.757 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.0001334 s, 3.8 MB/s 2026-03-21T06:48:40.757 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_3 2026-03-21T06:48:40.802 DEBUG:teuthology.orchestra.run.vm07:> stat /dev/vg_nvme/lv_4 2026-03-21T06:48:40.845 INFO:teuthology.orchestra.run.vm07.stdout: File: /dev/vg_nvme/lv_4 -> ../dm-3 2026-03-21T06:48:40.845 INFO:teuthology.orchestra.run.vm07.stdout: Size: 7 Blocks: 0 IO Block: 4096 symbolic link 2026-03-21T06:48:40.845 INFO:teuthology.orchestra.run.vm07.stdout:Device: 5h/5d Inode: 886 Links: 1 2026-03-21T06:48:40.845 INFO:teuthology.orchestra.run.vm07.stdout:Access: (0777/lrwxrwxrwx) Uid: ( 0/ root) Gid: ( 0/ root) 2026-03-21T06:48:40.845 INFO:teuthology.orchestra.run.vm07.stdout:Access: 2026-03-21 06:46:18.213309000 +0000 2026-03-21T06:48:40.845 INFO:teuthology.orchestra.run.vm07.stdout:Modify: 2026-03-21 06:45:51.557309000 +0000 2026-03-21T06:48:40.845 INFO:teuthology.orchestra.run.vm07.stdout:Change: 2026-03-21 06:45:51.557309000 +0000 2026-03-21T06:48:40.845 INFO:teuthology.orchestra.run.vm07.stdout: Birth: - 2026-03-21T06:48:40.845 DEBUG:teuthology.orchestra.run.vm07:> sudo dd if=/dev/vg_nvme/lv_4 of=/dev/null count=1 2026-03-21T06:48:40.892 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records in 2026-03-21T06:48:40.892 INFO:teuthology.orchestra.run.vm07.stderr:1+0 records out 2026-03-21T06:48:40.892 INFO:teuthology.orchestra.run.vm07.stderr:512 bytes copied, 0.000150602 s, 3.4 MB/s 2026-03-21T06:48:40.893 DEBUG:teuthology.orchestra.run.vm07:> ! mount | grep -v devtmpfs | grep -q /dev/vg_nvme/lv_4 2026-03-21T06:48:40.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:40 vm02 bash[17657]: cluster 2026-03-21T06:48:39.538735+0000 mgr.x (mgr.14152) 58 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:40.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:40 vm02 bash[17657]: cluster 2026-03-21T06:48:39.538735+0000 mgr.x (mgr.14152) 58 : cluster [DBG] pgmap v9: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:40.938 INFO:tasks.cephadm:Deploying osd.0 on vm02 with /dev/vg_nvme/lv_4... 2026-03-21T06:48:40.938 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- lvm zap /dev/vg_nvme/lv_4 2026-03-21T06:48:41.169 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:48:41.660 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:48:41.671 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch daemon add osd vm02:vg_nvme/lv_4 --skip-validation 2026-03-21T06:48:41.899 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: cluster 2026-03-21T06:48:41.538922+0000 mgr.x (mgr.14152) 59 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: cluster 2026-03-21T06:48:41.538922+0000 mgr.x (mgr.14152) 59 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.213474+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.213474+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.216969+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.216969+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.217887+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.217887+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.218022+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.218022+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.219816+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.219816+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.221831+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.221831+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.225007+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.225007+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.230823+0000 mon.a (mon.0) 268 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.230823+0000 mon.a (mon.0) 268 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.231564+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.231564+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.235525+0000 mon.a (mon.0) 270 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.235525+0000 mon.a (mon.0) 270 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.239082+0000 mon.a (mon.0) 271 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.239082+0000 mon.a (mon.0) 271 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.240068+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.240068+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.241243+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.241243+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.241700+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:42.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:42 vm02 bash[17657]: audit 2026-03-21T06:48:42.241700+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: cluster 2026-03-21T06:48:41.538922+0000 mgr.x (mgr.14152) 59 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: cluster 2026-03-21T06:48:41.538922+0000 mgr.x (mgr.14152) 59 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.213474+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.213474+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.216969+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.216969+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.217887+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.217887+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.218022+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.218022+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.219816+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.219816+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.221831+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.221831+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.225007+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.225007+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.230823+0000 mon.a (mon.0) 268 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.230823+0000 mon.a (mon.0) 268 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.231564+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.231564+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.235525+0000 mon.a (mon.0) 270 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.235525+0000 mon.a (mon.0) 270 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.239082+0000 mon.a (mon.0) 271 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.239082+0000 mon.a (mon.0) 271 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.240068+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.240068+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.241243+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.241243+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.241700+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:42.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:42 vm07 bash[19945]: audit 2026-03-21T06:48:42.241700+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:43.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: cluster 2026-03-21T06:48:41.538922+0000 mgr.x (mgr.14152) 59 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: cluster 2026-03-21T06:48:41.538922+0000 mgr.x (mgr.14152) 59 : cluster [DBG] pgmap v10: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.213474+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.213474+0000 mon.a (mon.0) 261 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.216969+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.216969+0000 mon.a (mon.0) 262 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.217887+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.217887+0000 mon.a (mon.0) 263 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.218022+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.218022+0000 mon.a (mon.0) 264 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.219816+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.219816+0000 mon.a (mon.0) 265 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.221831+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.221831+0000 mon.a (mon.0) 266 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.225007+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.225007+0000 mon.a (mon.0) 267 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.230823+0000 mon.a (mon.0) 268 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.230823+0000 mon.a (mon.0) 268 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.231564+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.231564+0000 mon.a (mon.0) 269 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.235525+0000 mon.a (mon.0) 270 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.235525+0000 mon.a (mon.0) 270 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.239082+0000 mon.a (mon.0) 271 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.239082+0000 mon.a (mon.0) 271 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.240068+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.240068+0000 mon.a (mon.0) 272 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.241243+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.241243+0000 mon.a (mon.0) 273 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.241700+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:42 vm04 bash[20194]: audit 2026-03-21T06:48:42.241700+0000 mon.a (mon.0) 274 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: audit 2026-03-21T06:48:42.208098+0000 mgr.x (mgr.14152) 60 : audit [DBG] from='client.14226 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: audit 2026-03-21T06:48:42.208098+0000 mgr.x (mgr.14152) 60 : audit [DBG] from='client.14226 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: cephadm 2026-03-21T06:48:42.208955+0000 mgr.x (mgr.14152) 61 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: cephadm 2026-03-21T06:48:42.208955+0000 mgr.x (mgr.14152) 61 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: cephadm 2026-03-21T06:48:42.209031+0000 mgr.x (mgr.14152) 62 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_4'] 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: cephadm 2026-03-21T06:48:42.209031+0000 mgr.x (mgr.14152) 62 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_4'] 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: cephadm 2026-03-21T06:48:42.213932+0000 mgr.x (mgr.14152) 63 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: cephadm 2026-03-21T06:48:42.213932+0000 mgr.x (mgr.14152) 63 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: cephadm 2026-03-21T06:48:42.214086+0000 mgr.x (mgr.14152) 64 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: cephadm 2026-03-21T06:48:42.214086+0000 mgr.x (mgr.14152) 64 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: audit 2026-03-21T06:48:43.102754+0000 mon.a (mon.0) 275 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"} : dispatch 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: audit 2026-03-21T06:48:43.102754+0000 mon.a (mon.0) 275 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"} : dispatch 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: audit 2026-03-21T06:48:43.105769+0000 mon.a (mon.0) 276 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"}]': finished 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: audit 2026-03-21T06:48:43.105769+0000 mon.a (mon.0) 276 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"}]': finished 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: cluster 2026-03-21T06:48:43.108139+0000 mon.a (mon.0) 277 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: cluster 2026-03-21T06:48:43.108139+0000 mon.a (mon.0) 277 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: audit 2026-03-21T06:48:43.108328+0000 mon.a (mon.0) 278 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: audit 2026-03-21T06:48:43.108328+0000 mon.a (mon.0) 278 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: audit 2026-03-21T06:48:43.445712+0000 mon.a (mon.0) 279 : audit [DBG] from='client.? 192.168.123.102:0/2488879623' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:48:43.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:43 vm02 bash[17657]: audit 2026-03-21T06:48:43.445712+0000 mon.a (mon.0) 279 : audit [DBG] from='client.? 192.168.123.102:0/2488879623' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: audit 2026-03-21T06:48:42.208098+0000 mgr.x (mgr.14152) 60 : audit [DBG] from='client.14226 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: audit 2026-03-21T06:48:42.208098+0000 mgr.x (mgr.14152) 60 : audit [DBG] from='client.14226 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: cephadm 2026-03-21T06:48:42.208955+0000 mgr.x (mgr.14152) 61 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: cephadm 2026-03-21T06:48:42.208955+0000 mgr.x (mgr.14152) 61 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: cephadm 2026-03-21T06:48:42.209031+0000 mgr.x (mgr.14152) 62 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_4'] 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: cephadm 2026-03-21T06:48:42.209031+0000 mgr.x (mgr.14152) 62 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_4'] 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: cephadm 2026-03-21T06:48:42.213932+0000 mgr.x (mgr.14152) 63 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: cephadm 2026-03-21T06:48:42.213932+0000 mgr.x (mgr.14152) 63 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: cephadm 2026-03-21T06:48:42.214086+0000 mgr.x (mgr.14152) 64 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: cephadm 2026-03-21T06:48:42.214086+0000 mgr.x (mgr.14152) 64 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: audit 2026-03-21T06:48:43.102754+0000 mon.a (mon.0) 275 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"} : dispatch 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: audit 2026-03-21T06:48:43.102754+0000 mon.a (mon.0) 275 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"} : dispatch 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: audit 2026-03-21T06:48:43.105769+0000 mon.a (mon.0) 276 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"}]': finished 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: audit 2026-03-21T06:48:43.105769+0000 mon.a (mon.0) 276 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"}]': finished 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: cluster 2026-03-21T06:48:43.108139+0000 mon.a (mon.0) 277 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: cluster 2026-03-21T06:48:43.108139+0000 mon.a (mon.0) 277 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: audit 2026-03-21T06:48:43.108328+0000 mon.a (mon.0) 278 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: audit 2026-03-21T06:48:43.108328+0000 mon.a (mon.0) 278 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: audit 2026-03-21T06:48:43.445712+0000 mon.a (mon.0) 279 : audit [DBG] from='client.? 192.168.123.102:0/2488879623' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:48:43.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:43 vm07 bash[19945]: audit 2026-03-21T06:48:43.445712+0000 mon.a (mon.0) 279 : audit [DBG] from='client.? 192.168.123.102:0/2488879623' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:48:44.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: audit 2026-03-21T06:48:42.208098+0000 mgr.x (mgr.14152) 60 : audit [DBG] from='client.14226 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:44.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: audit 2026-03-21T06:48:42.208098+0000 mgr.x (mgr.14152) 60 : audit [DBG] from='client.14226 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:48:44.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: cephadm 2026-03-21T06:48:42.208955+0000 mgr.x (mgr.14152) 61 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:48:44.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: cephadm 2026-03-21T06:48:42.208955+0000 mgr.x (mgr.14152) 61 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:48:44.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: cephadm 2026-03-21T06:48:42.209031+0000 mgr.x (mgr.14152) 62 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_4'] 2026-03-21T06:48:44.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: cephadm 2026-03-21T06:48:42.209031+0000 mgr.x (mgr.14152) 62 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_4'] 2026-03-21T06:48:44.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: cephadm 2026-03-21T06:48:42.213932+0000 mgr.x (mgr.14152) 63 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:48:44.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: cephadm 2026-03-21T06:48:42.213932+0000 mgr.x (mgr.14152) 63 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:48:44.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: cephadm 2026-03-21T06:48:42.214086+0000 mgr.x (mgr.14152) 64 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:48:44.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: cephadm 2026-03-21T06:48:42.214086+0000 mgr.x (mgr.14152) 64 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:48:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: audit 2026-03-21T06:48:43.102754+0000 mon.a (mon.0) 275 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"} : dispatch 2026-03-21T06:48:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: audit 2026-03-21T06:48:43.102754+0000 mon.a (mon.0) 275 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"} : dispatch 2026-03-21T06:48:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: audit 2026-03-21T06:48:43.105769+0000 mon.a (mon.0) 276 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"}]': finished 2026-03-21T06:48:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: audit 2026-03-21T06:48:43.105769+0000 mon.a (mon.0) 276 : audit [INF] from='client.? 192.168.123.102:0/1730313974' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "5a293892-6a20-454b-9116-106cddf14695"}]': finished 2026-03-21T06:48:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: cluster 2026-03-21T06:48:43.108139+0000 mon.a (mon.0) 277 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-21T06:48:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: cluster 2026-03-21T06:48:43.108139+0000 mon.a (mon.0) 277 : cluster [DBG] osdmap e5: 1 total, 0 up, 1 in 2026-03-21T06:48:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: audit 2026-03-21T06:48:43.108328+0000 mon.a (mon.0) 278 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: audit 2026-03-21T06:48:43.108328+0000 mon.a (mon.0) 278 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: audit 2026-03-21T06:48:43.445712+0000 mon.a (mon.0) 279 : audit [DBG] from='client.? 192.168.123.102:0/2488879623' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:48:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:43 vm04 bash[20194]: audit 2026-03-21T06:48:43.445712+0000 mon.a (mon.0) 279 : audit [DBG] from='client.? 192.168.123.102:0/2488879623' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:48:44.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:44 vm02 bash[17657]: cluster 2026-03-21T06:48:43.539102+0000 mgr.x (mgr.14152) 65 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:44.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:44 vm02 bash[17657]: cluster 2026-03-21T06:48:43.539102+0000 mgr.x (mgr.14152) 65 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:44.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:44 vm07 bash[19945]: cluster 2026-03-21T06:48:43.539102+0000 mgr.x (mgr.14152) 65 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:44.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:44 vm07 bash[19945]: cluster 2026-03-21T06:48:43.539102+0000 mgr.x (mgr.14152) 65 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:45.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:44 vm04 bash[20194]: cluster 2026-03-21T06:48:43.539102+0000 mgr.x (mgr.14152) 65 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:45.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:44 vm04 bash[20194]: cluster 2026-03-21T06:48:43.539102+0000 mgr.x (mgr.14152) 65 : cluster [DBG] pgmap v12: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:45.810 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:45 vm02 bash[17657]: audit 2026-03-21T06:48:45.237029+0000 mon.a (mon.0) 280 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:45.810 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:45 vm02 bash[17657]: audit 2026-03-21T06:48:45.237029+0000 mon.a (mon.0) 280 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:45.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:45 vm02 bash[17657]: audit 2026-03-21T06:48:45.237569+0000 mon.a (mon.0) 281 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:45.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:45 vm02 bash[17657]: audit 2026-03-21T06:48:45.237569+0000 mon.a (mon.0) 281 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:45.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:45 vm02 bash[17657]: audit 2026-03-21T06:48:45.549730+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:45.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:45 vm02 bash[17657]: audit 2026-03-21T06:48:45.549730+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:45.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:45 vm02 bash[17657]: audit 2026-03-21T06:48:45.551629+0000 mon.a (mon.0) 283 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:45.811 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:45 vm02 bash[17657]: audit 2026-03-21T06:48:45.551629+0000 mon.a (mon.0) 283 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:45.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:45 vm07 bash[19945]: audit 2026-03-21T06:48:45.237029+0000 mon.a (mon.0) 280 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:45.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:45 vm07 bash[19945]: audit 2026-03-21T06:48:45.237029+0000 mon.a (mon.0) 280 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:45.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:45 vm07 bash[19945]: audit 2026-03-21T06:48:45.237569+0000 mon.a (mon.0) 281 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:45.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:45 vm07 bash[19945]: audit 2026-03-21T06:48:45.237569+0000 mon.a (mon.0) 281 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:45.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:45 vm07 bash[19945]: audit 2026-03-21T06:48:45.549730+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:45.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:45 vm07 bash[19945]: audit 2026-03-21T06:48:45.549730+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:45.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:45 vm07 bash[19945]: audit 2026-03-21T06:48:45.551629+0000 mon.a (mon.0) 283 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:45.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:45 vm07 bash[19945]: audit 2026-03-21T06:48:45.551629+0000 mon.a (mon.0) 283 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:46.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:45 vm04 bash[20194]: audit 2026-03-21T06:48:45.237029+0000 mon.a (mon.0) 280 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:46.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:45 vm04 bash[20194]: audit 2026-03-21T06:48:45.237029+0000 mon.a (mon.0) 280 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:46.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:45 vm04 bash[20194]: audit 2026-03-21T06:48:45.237569+0000 mon.a (mon.0) 281 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:46.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:45 vm04 bash[20194]: audit 2026-03-21T06:48:45.237569+0000 mon.a (mon.0) 281 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:46.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:45 vm04 bash[20194]: audit 2026-03-21T06:48:45.549730+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:46.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:45 vm04 bash[20194]: audit 2026-03-21T06:48:45.549730+0000 mon.a (mon.0) 282 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.0"} : dispatch 2026-03-21T06:48:46.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:45 vm04 bash[20194]: audit 2026-03-21T06:48:45.551629+0000 mon.a (mon.0) 283 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:46.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:45 vm04 bash[20194]: audit 2026-03-21T06:48:45.551629+0000 mon.a (mon.0) 283 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:48:46.384 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:46.384 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:46.385 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:46 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:46.385 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:46 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:46.648 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 bash[17657]: cephadm 2026-03-21T06:48:45.237973+0000 mgr.x (mgr.14152) 66 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:46.648 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 bash[17657]: cephadm 2026-03-21T06:48:45.237973+0000 mgr.x (mgr.14152) 66 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:46.648 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 bash[17657]: cluster 2026-03-21T06:48:45.539281+0000 mgr.x (mgr.14152) 67 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:46.648 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 bash[17657]: cluster 2026-03-21T06:48:45.539281+0000 mgr.x (mgr.14152) 67 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:46.648 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 bash[17657]: cephadm 2026-03-21T06:48:45.552091+0000 mgr.x (mgr.14152) 68 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:46.648 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 bash[17657]: cephadm 2026-03-21T06:48:45.552091+0000 mgr.x (mgr.14152) 68 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:46.648 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 bash[17657]: audit 2026-03-21T06:48:46.417556+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:46.649 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 bash[17657]: audit 2026-03-21T06:48:46.417556+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:46.649 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 bash[17657]: audit 2026-03-21T06:48:46.427326+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:46.649 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:46 vm02 bash[17657]: audit 2026-03-21T06:48:46.427326+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:46 vm07 bash[19945]: cephadm 2026-03-21T06:48:45.237973+0000 mgr.x (mgr.14152) 66 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:46 vm07 bash[19945]: cephadm 2026-03-21T06:48:45.237973+0000 mgr.x (mgr.14152) 66 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:46 vm07 bash[19945]: cluster 2026-03-21T06:48:45.539281+0000 mgr.x (mgr.14152) 67 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:46 vm07 bash[19945]: cluster 2026-03-21T06:48:45.539281+0000 mgr.x (mgr.14152) 67 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:46 vm07 bash[19945]: cephadm 2026-03-21T06:48:45.552091+0000 mgr.x (mgr.14152) 68 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:46 vm07 bash[19945]: cephadm 2026-03-21T06:48:45.552091+0000 mgr.x (mgr.14152) 68 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:46 vm07 bash[19945]: audit 2026-03-21T06:48:46.417556+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:46 vm07 bash[19945]: audit 2026-03-21T06:48:46.417556+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:46 vm07 bash[19945]: audit 2026-03-21T06:48:46.427326+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:46 vm07 bash[19945]: audit 2026-03-21T06:48:46.427326+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:46 vm04 bash[20194]: cephadm 2026-03-21T06:48:45.237973+0000 mgr.x (mgr.14152) 66 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:46 vm04 bash[20194]: cephadm 2026-03-21T06:48:45.237973+0000 mgr.x (mgr.14152) 66 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:46 vm04 bash[20194]: cluster 2026-03-21T06:48:45.539281+0000 mgr.x (mgr.14152) 67 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:46 vm04 bash[20194]: cluster 2026-03-21T06:48:45.539281+0000 mgr.x (mgr.14152) 67 : cluster [DBG] pgmap v13: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:46 vm04 bash[20194]: cephadm 2026-03-21T06:48:45.552091+0000 mgr.x (mgr.14152) 68 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:46 vm04 bash[20194]: cephadm 2026-03-21T06:48:45.552091+0000 mgr.x (mgr.14152) 68 : cephadm [INF] Deploying daemon osd.0 on vm02 2026-03-21T06:48:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:46 vm04 bash[20194]: audit 2026-03-21T06:48:46.417556+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:46 vm04 bash[20194]: audit 2026-03-21T06:48:46.417556+0000 mon.a (mon.0) 284 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:46 vm04 bash[20194]: audit 2026-03-21T06:48:46.427326+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:47.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:46 vm04 bash[20194]: audit 2026-03-21T06:48:46.427326+0000 mon.a (mon.0) 285 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:48:47.207 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:48:47 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:47.207 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:47 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:48:48.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:48 vm02 bash[17657]: cluster 2026-03-21T06:48:47.539498+0000 mgr.x (mgr.14152) 69 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:48.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:48 vm02 bash[17657]: cluster 2026-03-21T06:48:47.539498+0000 mgr.x (mgr.14152) 69 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:48 vm07 bash[19945]: cluster 2026-03-21T06:48:47.539498+0000 mgr.x (mgr.14152) 69 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:48 vm07 bash[19945]: cluster 2026-03-21T06:48:47.539498+0000 mgr.x (mgr.14152) 69 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:49.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:48 vm04 bash[20194]: cluster 2026-03-21T06:48:47.539498+0000 mgr.x (mgr.14152) 69 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:49.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:48 vm04 bash[20194]: cluster 2026-03-21T06:48:47.539498+0000 mgr.x (mgr.14152) 69 : cluster [DBG] pgmap v14: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:49.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:49 vm02 bash[17657]: audit 2026-03-21T06:48:48.767611+0000 mon.a (mon.0) 286 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:48:49.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:49 vm02 bash[17657]: audit 2026-03-21T06:48:48.767611+0000 mon.a (mon.0) 286 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:48:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:49 vm07 bash[19945]: audit 2026-03-21T06:48:48.767611+0000 mon.a (mon.0) 286 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:48:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:49 vm07 bash[19945]: audit 2026-03-21T06:48:48.767611+0000 mon.a (mon.0) 286 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:48:50.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:49 vm04 bash[20194]: audit 2026-03-21T06:48:48.767611+0000 mon.a (mon.0) 286 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:48:50.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:49 vm04 bash[20194]: audit 2026-03-21T06:48:48.767611+0000 mon.a (mon.0) 286 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:48:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:50 vm07 bash[19945]: cluster 2026-03-21T06:48:49.539795+0000 mgr.x (mgr.14152) 70 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:50.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:50 vm07 bash[19945]: cluster 2026-03-21T06:48:49.539795+0000 mgr.x (mgr.14152) 70 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:50.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:50 vm07 bash[19945]: audit 2026-03-21T06:48:49.633811+0000 mon.a (mon.0) 287 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:48:50.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:50 vm07 bash[19945]: audit 2026-03-21T06:48:49.633811+0000 mon.a (mon.0) 287 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:48:50.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:50 vm07 bash[19945]: cluster 2026-03-21T06:48:49.635961+0000 mon.a (mon.0) 288 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-21T06:48:50.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:50 vm07 bash[19945]: cluster 2026-03-21T06:48:49.635961+0000 mon.a (mon.0) 288 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-21T06:48:50.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:50 vm07 bash[19945]: audit 2026-03-21T06:48:49.636121+0000 mon.a (mon.0) 289 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:50.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:50 vm07 bash[19945]: audit 2026-03-21T06:48:49.636121+0000 mon.a (mon.0) 289 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:50.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:50 vm07 bash[19945]: audit 2026-03-21T06:48:49.636249+0000 mon.a (mon.0) 290 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:48:50.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:50 vm07 bash[19945]: audit 2026-03-21T06:48:49.636249+0000 mon.a (mon.0) 290 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:48:51.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:50 vm04 bash[20194]: cluster 2026-03-21T06:48:49.539795+0000 mgr.x (mgr.14152) 70 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:51.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:50 vm04 bash[20194]: cluster 2026-03-21T06:48:49.539795+0000 mgr.x (mgr.14152) 70 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:51.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:50 vm04 bash[20194]: audit 2026-03-21T06:48:49.633811+0000 mon.a (mon.0) 287 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:48:51.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:50 vm04 bash[20194]: audit 2026-03-21T06:48:49.633811+0000 mon.a (mon.0) 287 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:48:51.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:50 vm04 bash[20194]: cluster 2026-03-21T06:48:49.635961+0000 mon.a (mon.0) 288 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-21T06:48:51.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:50 vm04 bash[20194]: cluster 2026-03-21T06:48:49.635961+0000 mon.a (mon.0) 288 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-21T06:48:51.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:50 vm04 bash[20194]: audit 2026-03-21T06:48:49.636121+0000 mon.a (mon.0) 289 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:51.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:50 vm04 bash[20194]: audit 2026-03-21T06:48:49.636121+0000 mon.a (mon.0) 289 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:51.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:50 vm04 bash[20194]: audit 2026-03-21T06:48:49.636249+0000 mon.a (mon.0) 290 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:48:51.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:50 vm04 bash[20194]: audit 2026-03-21T06:48:49.636249+0000 mon.a (mon.0) 290 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:48:51.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:50 vm02 bash[17657]: cluster 2026-03-21T06:48:49.539795+0000 mgr.x (mgr.14152) 70 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:51.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:50 vm02 bash[17657]: cluster 2026-03-21T06:48:49.539795+0000 mgr.x (mgr.14152) 70 : cluster [DBG] pgmap v15: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:51.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:50 vm02 bash[17657]: audit 2026-03-21T06:48:49.633811+0000 mon.a (mon.0) 287 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:48:51.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:50 vm02 bash[17657]: audit 2026-03-21T06:48:49.633811+0000 mon.a (mon.0) 287 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:48:51.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:50 vm02 bash[17657]: cluster 2026-03-21T06:48:49.635961+0000 mon.a (mon.0) 288 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-21T06:48:51.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:50 vm02 bash[17657]: cluster 2026-03-21T06:48:49.635961+0000 mon.a (mon.0) 288 : cluster [DBG] osdmap e6: 1 total, 0 up, 1 in 2026-03-21T06:48:51.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:50 vm02 bash[17657]: audit 2026-03-21T06:48:49.636121+0000 mon.a (mon.0) 289 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:51.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:50 vm02 bash[17657]: audit 2026-03-21T06:48:49.636121+0000 mon.a (mon.0) 289 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:51.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:50 vm02 bash[17657]: audit 2026-03-21T06:48:49.636249+0000 mon.a (mon.0) 290 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:48:51.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:50 vm02 bash[17657]: audit 2026-03-21T06:48:49.636249+0000 mon.a (mon.0) 290 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:48:52.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:51 vm04 bash[20194]: audit 2026-03-21T06:48:50.636450+0000 mon.a (mon.0) 291 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:48:52.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:51 vm04 bash[20194]: audit 2026-03-21T06:48:50.636450+0000 mon.a (mon.0) 291 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:48:52.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:51 vm04 bash[20194]: cluster 2026-03-21T06:48:50.638986+0000 mon.a (mon.0) 292 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-21T06:48:52.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:51 vm04 bash[20194]: cluster 2026-03-21T06:48:50.638986+0000 mon.a (mon.0) 292 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-21T06:48:52.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:51 vm04 bash[20194]: audit 2026-03-21T06:48:50.640031+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:52.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:51 vm04 bash[20194]: audit 2026-03-21T06:48:50.640031+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:52.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:51 vm04 bash[20194]: audit 2026-03-21T06:48:50.646447+0000 mon.a (mon.0) 294 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:52.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:51 vm04 bash[20194]: audit 2026-03-21T06:48:50.646447+0000 mon.a (mon.0) 294 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:52.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:51 vm02 bash[17657]: audit 2026-03-21T06:48:50.636450+0000 mon.a (mon.0) 291 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:48:52.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:51 vm02 bash[17657]: audit 2026-03-21T06:48:50.636450+0000 mon.a (mon.0) 291 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:48:52.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:51 vm02 bash[17657]: cluster 2026-03-21T06:48:50.638986+0000 mon.a (mon.0) 292 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-21T06:48:52.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:51 vm02 bash[17657]: cluster 2026-03-21T06:48:50.638986+0000 mon.a (mon.0) 292 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-21T06:48:52.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:51 vm02 bash[17657]: audit 2026-03-21T06:48:50.640031+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:52.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:51 vm02 bash[17657]: audit 2026-03-21T06:48:50.640031+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:52.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:51 vm02 bash[17657]: audit 2026-03-21T06:48:50.646447+0000 mon.a (mon.0) 294 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:52.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:51 vm02 bash[17657]: audit 2026-03-21T06:48:50.646447+0000 mon.a (mon.0) 294 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:52.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:51 vm07 bash[19945]: audit 2026-03-21T06:48:50.636450+0000 mon.a (mon.0) 291 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:48:52.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:51 vm07 bash[19945]: audit 2026-03-21T06:48:50.636450+0000 mon.a (mon.0) 291 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248]' entity='osd.0' cmd='[{"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:48:52.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:51 vm07 bash[19945]: cluster 2026-03-21T06:48:50.638986+0000 mon.a (mon.0) 292 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-21T06:48:52.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:51 vm07 bash[19945]: cluster 2026-03-21T06:48:50.638986+0000 mon.a (mon.0) 292 : cluster [DBG] osdmap e7: 1 total, 0 up, 1 in 2026-03-21T06:48:52.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:51 vm07 bash[19945]: audit 2026-03-21T06:48:50.640031+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:52.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:51 vm07 bash[19945]: audit 2026-03-21T06:48:50.640031+0000 mon.a (mon.0) 293 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:52.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:51 vm07 bash[19945]: audit 2026-03-21T06:48:50.646447+0000 mon.a (mon.0) 294 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:52.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:51 vm07 bash[19945]: audit 2026-03-21T06:48:50.646447+0000 mon.a (mon.0) 294 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: cluster 2026-03-21T06:48:49.774443+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: cluster 2026-03-21T06:48:49.774443+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: cluster 2026-03-21T06:48:49.774462+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: cluster 2026-03-21T06:48:49.774462+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: cluster 2026-03-21T06:48:51.540068+0000 mgr.x (mgr.14152) 71 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: cluster 2026-03-21T06:48:51.540068+0000 mgr.x (mgr.14152) 71 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: audit 2026-03-21T06:48:51.643096+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: audit 2026-03-21T06:48:51.643096+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: cluster 2026-03-21T06:48:51.653869+0000 mon.a (mon.0) 296 : cluster [INF] osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248] boot 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: cluster 2026-03-21T06:48:51.653869+0000 mon.a (mon.0) 296 : cluster [INF] osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248] boot 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: cluster 2026-03-21T06:48:51.653904+0000 mon.a (mon.0) 297 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: cluster 2026-03-21T06:48:51.653904+0000 mon.a (mon.0) 297 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: audit 2026-03-21T06:48:51.654100+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:52 vm04 bash[20194]: audit 2026-03-21T06:48:51.654100+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: cluster 2026-03-21T06:48:49.774443+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: cluster 2026-03-21T06:48:49.774443+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: cluster 2026-03-21T06:48:49.774462+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: cluster 2026-03-21T06:48:49.774462+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: cluster 2026-03-21T06:48:51.540068+0000 mgr.x (mgr.14152) 71 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: cluster 2026-03-21T06:48:51.540068+0000 mgr.x (mgr.14152) 71 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: audit 2026-03-21T06:48:51.643096+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: audit 2026-03-21T06:48:51.643096+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: cluster 2026-03-21T06:48:51.653869+0000 mon.a (mon.0) 296 : cluster [INF] osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248] boot 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: cluster 2026-03-21T06:48:51.653869+0000 mon.a (mon.0) 296 : cluster [INF] osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248] boot 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: cluster 2026-03-21T06:48:51.653904+0000 mon.a (mon.0) 297 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: cluster 2026-03-21T06:48:51.653904+0000 mon.a (mon.0) 297 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: audit 2026-03-21T06:48:51.654100+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:52 vm02 bash[17657]: audit 2026-03-21T06:48:51.654100+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: cluster 2026-03-21T06:48:49.774443+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:48:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: cluster 2026-03-21T06:48:49.774443+0000 osd.0 (osd.0) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:48:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: cluster 2026-03-21T06:48:49.774462+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:48:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: cluster 2026-03-21T06:48:49.774462+0000 osd.0 (osd.0) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:48:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: cluster 2026-03-21T06:48:51.540068+0000 mgr.x (mgr.14152) 71 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: cluster 2026-03-21T06:48:51.540068+0000 mgr.x (mgr.14152) 71 : cluster [DBG] pgmap v18: 0 pgs: ; 0 B data, 0 B used, 0 B / 0 B avail 2026-03-21T06:48:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: audit 2026-03-21T06:48:51.643096+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: audit 2026-03-21T06:48:51.643096+0000 mon.a (mon.0) 295 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: cluster 2026-03-21T06:48:51.653869+0000 mon.a (mon.0) 296 : cluster [INF] osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248] boot 2026-03-21T06:48:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: cluster 2026-03-21T06:48:51.653869+0000 mon.a (mon.0) 296 : cluster [INF] osd.0 [v2:192.168.123.102:6802/3921084248,v1:192.168.123.102:6803/3921084248] boot 2026-03-21T06:48:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: cluster 2026-03-21T06:48:51.653904+0000 mon.a (mon.0) 297 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-21T06:48:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: cluster 2026-03-21T06:48:51.653904+0000 mon.a (mon.0) 297 : cluster [DBG] osdmap e8: 1 total, 1 up, 1 in 2026-03-21T06:48:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: audit 2026-03-21T06:48:51.654100+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:52 vm07 bash[19945]: audit 2026-03-21T06:48:51.654100+0000 mon.a (mon.0) 298 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:48:54.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:53 vm04 bash[20194]: cluster 2026-03-21T06:48:52.657269+0000 mon.a (mon.0) 299 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-21T06:48:54.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:53 vm04 bash[20194]: cluster 2026-03-21T06:48:52.657269+0000 mon.a (mon.0) 299 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-21T06:48:54.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:53 vm02 bash[17657]: cluster 2026-03-21T06:48:52.657269+0000 mon.a (mon.0) 299 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-21T06:48:54.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:53 vm02 bash[17657]: cluster 2026-03-21T06:48:52.657269+0000 mon.a (mon.0) 299 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-21T06:48:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:53 vm07 bash[19945]: cluster 2026-03-21T06:48:52.657269+0000 mon.a (mon.0) 299 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-21T06:48:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:53 vm07 bash[19945]: cluster 2026-03-21T06:48:52.657269+0000 mon.a (mon.0) 299 : cluster [DBG] osdmap e9: 1 total, 1 up, 1 in 2026-03-21T06:48:55.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:54 vm04 bash[20194]: cluster 2026-03-21T06:48:53.540425+0000 mgr.x (mgr.14152) 72 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:55.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:54 vm04 bash[20194]: cluster 2026-03-21T06:48:53.540425+0000 mgr.x (mgr.14152) 72 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:55.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:54 vm02 bash[17657]: cluster 2026-03-21T06:48:53.540425+0000 mgr.x (mgr.14152) 72 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:55.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:54 vm02 bash[17657]: cluster 2026-03-21T06:48:53.540425+0000 mgr.x (mgr.14152) 72 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:55.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:54 vm07 bash[19945]: cluster 2026-03-21T06:48:53.540425+0000 mgr.x (mgr.14152) 72 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:55.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:54 vm07 bash[19945]: cluster 2026-03-21T06:48:53.540425+0000 mgr.x (mgr.14152) 72 : cluster [DBG] pgmap v21: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:56.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:55 vm04 bash[20194]: cluster 2026-03-21T06:48:55.540851+0000 mgr.x (mgr.14152) 73 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:56.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:55 vm04 bash[20194]: cluster 2026-03-21T06:48:55.540851+0000 mgr.x (mgr.14152) 73 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:56.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:55 vm02 bash[17657]: cluster 2026-03-21T06:48:55.540851+0000 mgr.x (mgr.14152) 73 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:56.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:55 vm02 bash[17657]: cluster 2026-03-21T06:48:55.540851+0000 mgr.x (mgr.14152) 73 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:56.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:55 vm07 bash[19945]: cluster 2026-03-21T06:48:55.540851+0000 mgr.x (mgr.14152) 73 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:56.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:55 vm07 bash[19945]: cluster 2026-03-21T06:48:55.540851+0000 mgr.x (mgr.14152) 73 : cluster [DBG] pgmap v22: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:58 vm02 bash[17657]: cluster 2026-03-21T06:48:57.541224+0000 mgr.x (mgr.14152) 74 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:48:58 vm02 bash[17657]: cluster 2026-03-21T06:48:57.541224+0000 mgr.x (mgr.14152) 74 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:58 vm07 bash[19945]: cluster 2026-03-21T06:48:57.541224+0000 mgr.x (mgr.14152) 74 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:48:58 vm07 bash[19945]: cluster 2026-03-21T06:48:57.541224+0000 mgr.x (mgr.14152) 74 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:59.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:58 vm04 bash[20194]: cluster 2026-03-21T06:48:57.541224+0000 mgr.x (mgr.14152) 74 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:48:59.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:48:58 vm04 bash[20194]: cluster 2026-03-21T06:48:57.541224+0000 mgr.x (mgr.14152) 74 : cluster [DBG] pgmap v23: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:00.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:00 vm02 bash[17657]: cluster 2026-03-21T06:48:59.541527+0000 mgr.x (mgr.14152) 75 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:00.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:00 vm02 bash[17657]: cluster 2026-03-21T06:48:59.541527+0000 mgr.x (mgr.14152) 75 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:00 vm07 bash[19945]: cluster 2026-03-21T06:48:59.541527+0000 mgr.x (mgr.14152) 75 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:00 vm07 bash[19945]: cluster 2026-03-21T06:48:59.541527+0000 mgr.x (mgr.14152) 75 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:01.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:00 vm04 bash[20194]: cluster 2026-03-21T06:48:59.541527+0000 mgr.x (mgr.14152) 75 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:01.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:00 vm04 bash[20194]: cluster 2026-03-21T06:48:59.541527+0000 mgr.x (mgr.14152) 75 : cluster [DBG] pgmap v24: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:02.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:02 vm02 bash[17657]: cluster 2026-03-21T06:49:01.541835+0000 mgr.x (mgr.14152) 76 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:02.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:02 vm02 bash[17657]: cluster 2026-03-21T06:49:01.541835+0000 mgr.x (mgr.14152) 76 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:02 vm07 bash[19945]: cluster 2026-03-21T06:49:01.541835+0000 mgr.x (mgr.14152) 76 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:02 vm07 bash[19945]: cluster 2026-03-21T06:49:01.541835+0000 mgr.x (mgr.14152) 76 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:03.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:02 vm04 bash[20194]: cluster 2026-03-21T06:49:01.541835+0000 mgr.x (mgr.14152) 76 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:03.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:02 vm04 bash[20194]: cluster 2026-03-21T06:49:01.541835+0000 mgr.x (mgr.14152) 76 : cluster [DBG] pgmap v25: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:04 vm02 bash[17657]: cluster 2026-03-21T06:49:03.542214+0000 mgr.x (mgr.14152) 77 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:04 vm02 bash[17657]: cluster 2026-03-21T06:49:03.542214+0000 mgr.x (mgr.14152) 77 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:04 vm07 bash[19945]: cluster 2026-03-21T06:49:03.542214+0000 mgr.x (mgr.14152) 77 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:04 vm07 bash[19945]: cluster 2026-03-21T06:49:03.542214+0000 mgr.x (mgr.14152) 77 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:05.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:04 vm04 bash[20194]: cluster 2026-03-21T06:49:03.542214+0000 mgr.x (mgr.14152) 77 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:05.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:04 vm04 bash[20194]: cluster 2026-03-21T06:49:03.542214+0000 mgr.x (mgr.14152) 77 : cluster [DBG] pgmap v26: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:06 vm02 bash[17657]: cluster 2026-03-21T06:49:05.542578+0000 mgr.x (mgr.14152) 78 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:06 vm02 bash[17657]: cluster 2026-03-21T06:49:05.542578+0000 mgr.x (mgr.14152) 78 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:06 vm07 bash[19945]: cluster 2026-03-21T06:49:05.542578+0000 mgr.x (mgr.14152) 78 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:06 vm07 bash[19945]: cluster 2026-03-21T06:49:05.542578+0000 mgr.x (mgr.14152) 78 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:06 vm04 bash[20194]: cluster 2026-03-21T06:49:05.542578+0000 mgr.x (mgr.14152) 78 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:06 vm04 bash[20194]: cluster 2026-03-21T06:49:05.542578+0000 mgr.x (mgr.14152) 78 : cluster [DBG] pgmap v27: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:08.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:08 vm02 bash[17657]: cluster 2026-03-21T06:49:07.542948+0000 mgr.x (mgr.14152) 79 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:08.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:08 vm02 bash[17657]: cluster 2026-03-21T06:49:07.542948+0000 mgr.x (mgr.14152) 79 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:08.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:08 vm07 bash[19945]: cluster 2026-03-21T06:49:07.542948+0000 mgr.x (mgr.14152) 79 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:08.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:08 vm07 bash[19945]: cluster 2026-03-21T06:49:07.542948+0000 mgr.x (mgr.14152) 79 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:09.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:08 vm04 bash[20194]: cluster 2026-03-21T06:49:07.542948+0000 mgr.x (mgr.14152) 79 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:09.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:08 vm04 bash[20194]: cluster 2026-03-21T06:49:07.542948+0000 mgr.x (mgr.14152) 79 : cluster [DBG] pgmap v28: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:10.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:10 vm02 bash[17657]: cluster 2026-03-21T06:49:09.543225+0000 mgr.x (mgr.14152) 80 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:10.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:10 vm02 bash[17657]: cluster 2026-03-21T06:49:09.543225+0000 mgr.x (mgr.14152) 80 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:10.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:10 vm07 bash[19945]: cluster 2026-03-21T06:49:09.543225+0000 mgr.x (mgr.14152) 80 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:10.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:10 vm07 bash[19945]: cluster 2026-03-21T06:49:09.543225+0000 mgr.x (mgr.14152) 80 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:11.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:10 vm04 bash[20194]: cluster 2026-03-21T06:49:09.543225+0000 mgr.x (mgr.14152) 80 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:11.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:10 vm04 bash[20194]: cluster 2026-03-21T06:49:09.543225+0000 mgr.x (mgr.14152) 80 : cluster [DBG] pgmap v29: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:12.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:12 vm07 bash[19945]: cluster 2026-03-21T06:49:11.543549+0000 mgr.x (mgr.14152) 81 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:12.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:12 vm07 bash[19945]: cluster 2026-03-21T06:49:11.543549+0000 mgr.x (mgr.14152) 81 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:13.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:12 vm04 bash[20194]: cluster 2026-03-21T06:49:11.543549+0000 mgr.x (mgr.14152) 81 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:13.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:12 vm04 bash[20194]: cluster 2026-03-21T06:49:11.543549+0000 mgr.x (mgr.14152) 81 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:12 vm02 bash[17657]: cluster 2026-03-21T06:49:11.543549+0000 mgr.x (mgr.14152) 81 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:12 vm02 bash[17657]: cluster 2026-03-21T06:49:11.543549+0000 mgr.x (mgr.14152) 81 : cluster [DBG] pgmap v30: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:15.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:14 vm04 bash[20194]: cluster 2026-03-21T06:49:13.543838+0000 mgr.x (mgr.14152) 82 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:15.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:14 vm04 bash[20194]: cluster 2026-03-21T06:49:13.543838+0000 mgr.x (mgr.14152) 82 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:15.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:14 vm02 bash[17657]: cluster 2026-03-21T06:49:13.543838+0000 mgr.x (mgr.14152) 82 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:15.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:14 vm02 bash[17657]: cluster 2026-03-21T06:49:13.543838+0000 mgr.x (mgr.14152) 82 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:14 vm07 bash[19945]: cluster 2026-03-21T06:49:13.543838+0000 mgr.x (mgr.14152) 82 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:14 vm07 bash[19945]: cluster 2026-03-21T06:49:13.543838+0000 mgr.x (mgr.14152) 82 : cluster [DBG] pgmap v31: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:17.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:16 vm04 bash[20194]: cluster 2026-03-21T06:49:15.544068+0000 mgr.x (mgr.14152) 83 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:17.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:16 vm04 bash[20194]: cluster 2026-03-21T06:49:15.544068+0000 mgr.x (mgr.14152) 83 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:17.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:16 vm02 bash[17657]: cluster 2026-03-21T06:49:15.544068+0000 mgr.x (mgr.14152) 83 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:17.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:16 vm02 bash[17657]: cluster 2026-03-21T06:49:15.544068+0000 mgr.x (mgr.14152) 83 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:16 vm07 bash[19945]: cluster 2026-03-21T06:49:15.544068+0000 mgr.x (mgr.14152) 83 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:16 vm07 bash[19945]: cluster 2026-03-21T06:49:15.544068+0000 mgr.x (mgr.14152) 83 : cluster [DBG] pgmap v32: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:19.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:18 vm04 bash[20194]: cluster 2026-03-21T06:49:17.544299+0000 mgr.x (mgr.14152) 84 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:19.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:18 vm04 bash[20194]: cluster 2026-03-21T06:49:17.544299+0000 mgr.x (mgr.14152) 84 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:18 vm02 bash[17657]: cluster 2026-03-21T06:49:17.544299+0000 mgr.x (mgr.14152) 84 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:18 vm02 bash[17657]: cluster 2026-03-21T06:49:17.544299+0000 mgr.x (mgr.14152) 84 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:18 vm07 bash[19945]: cluster 2026-03-21T06:49:17.544299+0000 mgr.x (mgr.14152) 84 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:18 vm07 bash[19945]: cluster 2026-03-21T06:49:17.544299+0000 mgr.x (mgr.14152) 84 : cluster [DBG] pgmap v33: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:21.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:20 vm04 bash[20194]: cluster 2026-03-21T06:49:19.544515+0000 mgr.x (mgr.14152) 85 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:21.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:20 vm04 bash[20194]: cluster 2026-03-21T06:49:19.544515+0000 mgr.x (mgr.14152) 85 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:21.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:20 vm02 bash[17657]: cluster 2026-03-21T06:49:19.544515+0000 mgr.x (mgr.14152) 85 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:21.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:20 vm02 bash[17657]: cluster 2026-03-21T06:49:19.544515+0000 mgr.x (mgr.14152) 85 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:20 vm07 bash[19945]: cluster 2026-03-21T06:49:19.544515+0000 mgr.x (mgr.14152) 85 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:20 vm07 bash[19945]: cluster 2026-03-21T06:49:19.544515+0000 mgr.x (mgr.14152) 85 : cluster [DBG] pgmap v34: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:22.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:21 vm04 bash[20194]: cluster 2026-03-21T06:49:21.544850+0000 mgr.x (mgr.14152) 86 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:22.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:21 vm04 bash[20194]: cluster 2026-03-21T06:49:21.544850+0000 mgr.x (mgr.14152) 86 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:22.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:21 vm02 bash[17657]: cluster 2026-03-21T06:49:21.544850+0000 mgr.x (mgr.14152) 86 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:22.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:21 vm02 bash[17657]: cluster 2026-03-21T06:49:21.544850+0000 mgr.x (mgr.14152) 86 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:22.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:21 vm07 bash[19945]: cluster 2026-03-21T06:49:21.544850+0000 mgr.x (mgr.14152) 86 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:22.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:21 vm07 bash[19945]: cluster 2026-03-21T06:49:21.544850+0000 mgr.x (mgr.14152) 86 : cluster [DBG] pgmap v35: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:24 vm02 bash[17657]: cluster 2026-03-21T06:49:23.545138+0000 mgr.x (mgr.14152) 87 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:24 vm02 bash[17657]: cluster 2026-03-21T06:49:23.545138+0000 mgr.x (mgr.14152) 87 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:24.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:24 vm07 bash[19945]: cluster 2026-03-21T06:49:23.545138+0000 mgr.x (mgr.14152) 87 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:24.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:24 vm07 bash[19945]: cluster 2026-03-21T06:49:23.545138+0000 mgr.x (mgr.14152) 87 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:25.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:24 vm04 bash[20194]: cluster 2026-03-21T06:49:23.545138+0000 mgr.x (mgr.14152) 87 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:25.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:24 vm04 bash[20194]: cluster 2026-03-21T06:49:23.545138+0000 mgr.x (mgr.14152) 87 : cluster [DBG] pgmap v36: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:26.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:26 vm02 bash[17657]: cluster 2026-03-21T06:49:25.545406+0000 mgr.x (mgr.14152) 88 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:26.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:26 vm02 bash[17657]: cluster 2026-03-21T06:49:25.545406+0000 mgr.x (mgr.14152) 88 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:26.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:26 vm07 bash[19945]: cluster 2026-03-21T06:49:25.545406+0000 mgr.x (mgr.14152) 88 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:26.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:26 vm07 bash[19945]: cluster 2026-03-21T06:49:25.545406+0000 mgr.x (mgr.14152) 88 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:27.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:26 vm04 bash[20194]: cluster 2026-03-21T06:49:25.545406+0000 mgr.x (mgr.14152) 88 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:27.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:26 vm04 bash[20194]: cluster 2026-03-21T06:49:25.545406+0000 mgr.x (mgr.14152) 88 : cluster [DBG] pgmap v37: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:28.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:28 vm02 bash[17657]: cluster 2026-03-21T06:49:27.545678+0000 mgr.x (mgr.14152) 89 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:28.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:28 vm02 bash[17657]: cluster 2026-03-21T06:49:27.545678+0000 mgr.x (mgr.14152) 89 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:28.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:28 vm07 bash[19945]: cluster 2026-03-21T06:49:27.545678+0000 mgr.x (mgr.14152) 89 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:28.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:28 vm07 bash[19945]: cluster 2026-03-21T06:49:27.545678+0000 mgr.x (mgr.14152) 89 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:29.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:28 vm04 bash[20194]: cluster 2026-03-21T06:49:27.545678+0000 mgr.x (mgr.14152) 89 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:29.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:28 vm04 bash[20194]: cluster 2026-03-21T06:49:27.545678+0000 mgr.x (mgr.14152) 89 : cluster [DBG] pgmap v38: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:30 vm02 bash[17657]: cluster 2026-03-21T06:49:29.545941+0000 mgr.x (mgr.14152) 90 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:30 vm02 bash[17657]: cluster 2026-03-21T06:49:29.545941+0000 mgr.x (mgr.14152) 90 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:30 vm07 bash[19945]: cluster 2026-03-21T06:49:29.545941+0000 mgr.x (mgr.14152) 90 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:30 vm07 bash[19945]: cluster 2026-03-21T06:49:29.545941+0000 mgr.x (mgr.14152) 90 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:31.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:30 vm04 bash[20194]: cluster 2026-03-21T06:49:29.545941+0000 mgr.x (mgr.14152) 90 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:31.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:30 vm04 bash[20194]: cluster 2026-03-21T06:49:29.545941+0000 mgr.x (mgr.14152) 90 : cluster [DBG] pgmap v39: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:32.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:32 vm02 bash[17657]: cluster 2026-03-21T06:49:31.546188+0000 mgr.x (mgr.14152) 91 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:32.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:32 vm02 bash[17657]: cluster 2026-03-21T06:49:31.546188+0000 mgr.x (mgr.14152) 91 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:32.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:32 vm07 bash[19945]: cluster 2026-03-21T06:49:31.546188+0000 mgr.x (mgr.14152) 91 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:32.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:32 vm07 bash[19945]: cluster 2026-03-21T06:49:31.546188+0000 mgr.x (mgr.14152) 91 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:33.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:32 vm04 bash[20194]: cluster 2026-03-21T06:49:31.546188+0000 mgr.x (mgr.14152) 91 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:33.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:32 vm04 bash[20194]: cluster 2026-03-21T06:49:31.546188+0000 mgr.x (mgr.14152) 91 : cluster [DBG] pgmap v40: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:34.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:34 vm02 bash[17657]: cluster 2026-03-21T06:49:33.546480+0000 mgr.x (mgr.14152) 92 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:34.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:34 vm02 bash[17657]: cluster 2026-03-21T06:49:33.546480+0000 mgr.x (mgr.14152) 92 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:34.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:34 vm07 bash[19945]: cluster 2026-03-21T06:49:33.546480+0000 mgr.x (mgr.14152) 92 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:34.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:34 vm07 bash[19945]: cluster 2026-03-21T06:49:33.546480+0000 mgr.x (mgr.14152) 92 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:35.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:34 vm04 bash[20194]: cluster 2026-03-21T06:49:33.546480+0000 mgr.x (mgr.14152) 92 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:35.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:34 vm04 bash[20194]: cluster 2026-03-21T06:49:33.546480+0000 mgr.x (mgr.14152) 92 : cluster [DBG] pgmap v41: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:36 vm07 bash[19945]: cluster 2026-03-21T06:49:35.546689+0000 mgr.x (mgr.14152) 93 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:36 vm07 bash[19945]: cluster 2026-03-21T06:49:35.546689+0000 mgr.x (mgr.14152) 93 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:37.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:36 vm04 bash[20194]: cluster 2026-03-21T06:49:35.546689+0000 mgr.x (mgr.14152) 93 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:37.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:36 vm04 bash[20194]: cluster 2026-03-21T06:49:35.546689+0000 mgr.x (mgr.14152) 93 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:37.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:36 vm02 bash[17657]: cluster 2026-03-21T06:49:35.546689+0000 mgr.x (mgr.14152) 93 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:37.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:36 vm02 bash[17657]: cluster 2026-03-21T06:49:35.546689+0000 mgr.x (mgr.14152) 93 : cluster [DBG] pgmap v42: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:39.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:38 vm04 bash[20194]: cluster 2026-03-21T06:49:37.546931+0000 mgr.x (mgr.14152) 94 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:39.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:38 vm04 bash[20194]: cluster 2026-03-21T06:49:37.546931+0000 mgr.x (mgr.14152) 94 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:39.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:38 vm02 bash[17657]: cluster 2026-03-21T06:49:37.546931+0000 mgr.x (mgr.14152) 94 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:39.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:38 vm02 bash[17657]: cluster 2026-03-21T06:49:37.546931+0000 mgr.x (mgr.14152) 94 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:38 vm07 bash[19945]: cluster 2026-03-21T06:49:37.546931+0000 mgr.x (mgr.14152) 94 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:38 vm07 bash[19945]: cluster 2026-03-21T06:49:37.546931+0000 mgr.x (mgr.14152) 94 : cluster [DBG] pgmap v43: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:41.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:40 vm04 bash[20194]: cluster 2026-03-21T06:49:39.547221+0000 mgr.x (mgr.14152) 95 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:41.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:40 vm04 bash[20194]: cluster 2026-03-21T06:49:39.547221+0000 mgr.x (mgr.14152) 95 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:41.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:40 vm02 bash[17657]: cluster 2026-03-21T06:49:39.547221+0000 mgr.x (mgr.14152) 95 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:41.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:40 vm02 bash[17657]: cluster 2026-03-21T06:49:39.547221+0000 mgr.x (mgr.14152) 95 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:40 vm07 bash[19945]: cluster 2026-03-21T06:49:39.547221+0000 mgr.x (mgr.14152) 95 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:40 vm07 bash[19945]: cluster 2026-03-21T06:49:39.547221+0000 mgr.x (mgr.14152) 95 : cluster [DBG] pgmap v44: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:43.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:42 vm04 bash[20194]: cluster 2026-03-21T06:49:41.547462+0000 mgr.x (mgr.14152) 96 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:43.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:42 vm04 bash[20194]: cluster 2026-03-21T06:49:41.547462+0000 mgr.x (mgr.14152) 96 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:43.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:42 vm02 bash[17657]: cluster 2026-03-21T06:49:41.547462+0000 mgr.x (mgr.14152) 96 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:43.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:42 vm02 bash[17657]: cluster 2026-03-21T06:49:41.547462+0000 mgr.x (mgr.14152) 96 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:42 vm07 bash[19945]: cluster 2026-03-21T06:49:41.547462+0000 mgr.x (mgr.14152) 96 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:42 vm07 bash[19945]: cluster 2026-03-21T06:49:41.547462+0000 mgr.x (mgr.14152) 96 : cluster [DBG] pgmap v45: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:45.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:44 vm04 bash[20194]: cluster 2026-03-21T06:49:43.547722+0000 mgr.x (mgr.14152) 97 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:45.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:44 vm04 bash[20194]: cluster 2026-03-21T06:49:43.547722+0000 mgr.x (mgr.14152) 97 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:45.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:44 vm02 bash[17657]: cluster 2026-03-21T06:49:43.547722+0000 mgr.x (mgr.14152) 97 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:45.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:44 vm02 bash[17657]: cluster 2026-03-21T06:49:43.547722+0000 mgr.x (mgr.14152) 97 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:45.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:44 vm07 bash[19945]: cluster 2026-03-21T06:49:43.547722+0000 mgr.x (mgr.14152) 97 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:45.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:44 vm07 bash[19945]: cluster 2026-03-21T06:49:43.547722+0000 mgr.x (mgr.14152) 97 : cluster [DBG] pgmap v46: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:46 vm04 bash[20194]: cluster 2026-03-21T06:49:45.548012+0000 mgr.x (mgr.14152) 98 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:46 vm04 bash[20194]: cluster 2026-03-21T06:49:45.548012+0000 mgr.x (mgr.14152) 98 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:47.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:46 vm02 bash[17657]: cluster 2026-03-21T06:49:45.548012+0000 mgr.x (mgr.14152) 98 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:47.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:46 vm02 bash[17657]: cluster 2026-03-21T06:49:45.548012+0000 mgr.x (mgr.14152) 98 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:47.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:46 vm07 bash[19945]: cluster 2026-03-21T06:49:45.548012+0000 mgr.x (mgr.14152) 98 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:47.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:46 vm07 bash[19945]: cluster 2026-03-21T06:49:45.548012+0000 mgr.x (mgr.14152) 98 : cluster [DBG] pgmap v47: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:48.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:47 vm04 bash[20194]: cluster 2026-03-21T06:49:47.548311+0000 mgr.x (mgr.14152) 99 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:48.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:47 vm04 bash[20194]: cluster 2026-03-21T06:49:47.548311+0000 mgr.x (mgr.14152) 99 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:48.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:47 vm02 bash[17657]: cluster 2026-03-21T06:49:47.548311+0000 mgr.x (mgr.14152) 99 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:48.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:47 vm02 bash[17657]: cluster 2026-03-21T06:49:47.548311+0000 mgr.x (mgr.14152) 99 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:48.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:47 vm07 bash[19945]: cluster 2026-03-21T06:49:47.548311+0000 mgr.x (mgr.14152) 99 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:48.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:47 vm07 bash[19945]: cluster 2026-03-21T06:49:47.548311+0000 mgr.x (mgr.14152) 99 : cluster [DBG] pgmap v48: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:50.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:50 vm02 bash[17657]: cluster 2026-03-21T06:49:49.548538+0000 mgr.x (mgr.14152) 100 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:50.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:50 vm02 bash[17657]: cluster 2026-03-21T06:49:49.548538+0000 mgr.x (mgr.14152) 100 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:50 vm07 bash[19945]: cluster 2026-03-21T06:49:49.548538+0000 mgr.x (mgr.14152) 100 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:50 vm07 bash[19945]: cluster 2026-03-21T06:49:49.548538+0000 mgr.x (mgr.14152) 100 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:51.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:50 vm04 bash[20194]: cluster 2026-03-21T06:49:49.548538+0000 mgr.x (mgr.14152) 100 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:51.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:50 vm04 bash[20194]: cluster 2026-03-21T06:49:49.548538+0000 mgr.x (mgr.14152) 100 : cluster [DBG] pgmap v49: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:52 vm02 bash[17657]: cluster 2026-03-21T06:49:51.548781+0000 mgr.x (mgr.14152) 101 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:52 vm02 bash[17657]: cluster 2026-03-21T06:49:51.548781+0000 mgr.x (mgr.14152) 101 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:52.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:52 vm07 bash[19945]: cluster 2026-03-21T06:49:51.548781+0000 mgr.x (mgr.14152) 101 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:52.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:52 vm07 bash[19945]: cluster 2026-03-21T06:49:51.548781+0000 mgr.x (mgr.14152) 101 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:53.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:52 vm04 bash[20194]: cluster 2026-03-21T06:49:51.548781+0000 mgr.x (mgr.14152) 101 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:53.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:52 vm04 bash[20194]: cluster 2026-03-21T06:49:51.548781+0000 mgr.x (mgr.14152) 101 : cluster [DBG] pgmap v50: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:54.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:54 vm02 bash[17657]: cluster 2026-03-21T06:49:53.549043+0000 mgr.x (mgr.14152) 102 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:54.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:54 vm02 bash[17657]: cluster 2026-03-21T06:49:53.549043+0000 mgr.x (mgr.14152) 102 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:54.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:54 vm07 bash[19945]: cluster 2026-03-21T06:49:53.549043+0000 mgr.x (mgr.14152) 102 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:54.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:54 vm07 bash[19945]: cluster 2026-03-21T06:49:53.549043+0000 mgr.x (mgr.14152) 102 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:55.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:54 vm04 bash[20194]: cluster 2026-03-21T06:49:53.549043+0000 mgr.x (mgr.14152) 102 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:55.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:54 vm04 bash[20194]: cluster 2026-03-21T06:49:53.549043+0000 mgr.x (mgr.14152) 102 : cluster [DBG] pgmap v51: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:56.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:56 vm02 bash[17657]: cluster 2026-03-21T06:49:55.549280+0000 mgr.x (mgr.14152) 103 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:56.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:56 vm02 bash[17657]: cluster 2026-03-21T06:49:55.549280+0000 mgr.x (mgr.14152) 103 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:56.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:56 vm07 bash[19945]: cluster 2026-03-21T06:49:55.549280+0000 mgr.x (mgr.14152) 103 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:56.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:56 vm07 bash[19945]: cluster 2026-03-21T06:49:55.549280+0000 mgr.x (mgr.14152) 103 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:57.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:56 vm04 bash[20194]: cluster 2026-03-21T06:49:55.549280+0000 mgr.x (mgr.14152) 103 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:57.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:56 vm04 bash[20194]: cluster 2026-03-21T06:49:55.549280+0000 mgr.x (mgr.14152) 103 : cluster [DBG] pgmap v52: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:58 vm02 bash[17657]: cluster 2026-03-21T06:49:57.549536+0000 mgr.x (mgr.14152) 104 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:58.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:49:58 vm02 bash[17657]: cluster 2026-03-21T06:49:57.549536+0000 mgr.x (mgr.14152) 104 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:58 vm07 bash[19945]: cluster 2026-03-21T06:49:57.549536+0000 mgr.x (mgr.14152) 104 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:49:58 vm07 bash[19945]: cluster 2026-03-21T06:49:57.549536+0000 mgr.x (mgr.14152) 104 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:59.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:58 vm04 bash[20194]: cluster 2026-03-21T06:49:57.549536+0000 mgr.x (mgr.14152) 104 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:49:59.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:49:58 vm04 bash[20194]: cluster 2026-03-21T06:49:57.549536+0000 mgr.x (mgr.14152) 104 : cluster [DBG] pgmap v53: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:00.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:00 vm02 bash[17657]: cluster 2026-03-21T06:49:59.549766+0000 mgr.x (mgr.14152) 105 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:00.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:00 vm02 bash[17657]: cluster 2026-03-21T06:49:59.549766+0000 mgr.x (mgr.14152) 105 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:00.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:00 vm02 bash[17657]: cluster 2026-03-21T06:50:00.000134+0000 mon.a (mon.0) 300 : cluster [INF] overall HEALTH_OK 2026-03-21T06:50:00.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:00 vm02 bash[17657]: cluster 2026-03-21T06:50:00.000134+0000 mon.a (mon.0) 300 : cluster [INF] overall HEALTH_OK 2026-03-21T06:50:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:00 vm07 bash[19945]: cluster 2026-03-21T06:49:59.549766+0000 mgr.x (mgr.14152) 105 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:00 vm07 bash[19945]: cluster 2026-03-21T06:49:59.549766+0000 mgr.x (mgr.14152) 105 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:00 vm07 bash[19945]: cluster 2026-03-21T06:50:00.000134+0000 mon.a (mon.0) 300 : cluster [INF] overall HEALTH_OK 2026-03-21T06:50:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:00 vm07 bash[19945]: cluster 2026-03-21T06:50:00.000134+0000 mon.a (mon.0) 300 : cluster [INF] overall HEALTH_OK 2026-03-21T06:50:01.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:00 vm04 bash[20194]: cluster 2026-03-21T06:49:59.549766+0000 mgr.x (mgr.14152) 105 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:01.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:00 vm04 bash[20194]: cluster 2026-03-21T06:49:59.549766+0000 mgr.x (mgr.14152) 105 : cluster [DBG] pgmap v54: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:01.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:00 vm04 bash[20194]: cluster 2026-03-21T06:50:00.000134+0000 mon.a (mon.0) 300 : cluster [INF] overall HEALTH_OK 2026-03-21T06:50:01.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:00 vm04 bash[20194]: cluster 2026-03-21T06:50:00.000134+0000 mon.a (mon.0) 300 : cluster [INF] overall HEALTH_OK 2026-03-21T06:50:02.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:02 vm02 bash[17657]: cluster 2026-03-21T06:50:01.550050+0000 mgr.x (mgr.14152) 106 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:02.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:02 vm02 bash[17657]: cluster 2026-03-21T06:50:01.550050+0000 mgr.x (mgr.14152) 106 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:02 vm07 bash[19945]: cluster 2026-03-21T06:50:01.550050+0000 mgr.x (mgr.14152) 106 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:02 vm07 bash[19945]: cluster 2026-03-21T06:50:01.550050+0000 mgr.x (mgr.14152) 106 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:03.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:02 vm04 bash[20194]: cluster 2026-03-21T06:50:01.550050+0000 mgr.x (mgr.14152) 106 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:03.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:02 vm04 bash[20194]: cluster 2026-03-21T06:50:01.550050+0000 mgr.x (mgr.14152) 106 : cluster [DBG] pgmap v55: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:04 vm07 bash[19945]: cluster 2026-03-21T06:50:03.550373+0000 mgr.x (mgr.14152) 107 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:04 vm07 bash[19945]: cluster 2026-03-21T06:50:03.550373+0000 mgr.x (mgr.14152) 107 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:05.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:04 vm04 bash[20194]: cluster 2026-03-21T06:50:03.550373+0000 mgr.x (mgr.14152) 107 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:05.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:04 vm04 bash[20194]: cluster 2026-03-21T06:50:03.550373+0000 mgr.x (mgr.14152) 107 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:05.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:04 vm02 bash[17657]: cluster 2026-03-21T06:50:03.550373+0000 mgr.x (mgr.14152) 107 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:05.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:04 vm02 bash[17657]: cluster 2026-03-21T06:50:03.550373+0000 mgr.x (mgr.14152) 107 : cluster [DBG] pgmap v56: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:07.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:06 vm04 bash[20194]: cluster 2026-03-21T06:50:05.550597+0000 mgr.x (mgr.14152) 108 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:07.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:06 vm04 bash[20194]: cluster 2026-03-21T06:50:05.550597+0000 mgr.x (mgr.14152) 108 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:07.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:06 vm02 bash[17657]: cluster 2026-03-21T06:50:05.550597+0000 mgr.x (mgr.14152) 108 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:07.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:06 vm02 bash[17657]: cluster 2026-03-21T06:50:05.550597+0000 mgr.x (mgr.14152) 108 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:06 vm07 bash[19945]: cluster 2026-03-21T06:50:05.550597+0000 mgr.x (mgr.14152) 108 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:06 vm07 bash[19945]: cluster 2026-03-21T06:50:05.550597+0000 mgr.x (mgr.14152) 108 : cluster [DBG] pgmap v57: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:09.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:08 vm04 bash[20194]: cluster 2026-03-21T06:50:07.550854+0000 mgr.x (mgr.14152) 109 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:09.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:08 vm04 bash[20194]: cluster 2026-03-21T06:50:07.550854+0000 mgr.x (mgr.14152) 109 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:08 vm02 bash[17657]: cluster 2026-03-21T06:50:07.550854+0000 mgr.x (mgr.14152) 109 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:08 vm02 bash[17657]: cluster 2026-03-21T06:50:07.550854+0000 mgr.x (mgr.14152) 109 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:08 vm07 bash[19945]: cluster 2026-03-21T06:50:07.550854+0000 mgr.x (mgr.14152) 109 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:08 vm07 bash[19945]: cluster 2026-03-21T06:50:07.550854+0000 mgr.x (mgr.14152) 109 : cluster [DBG] pgmap v58: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:11.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:10 vm04 bash[20194]: cluster 2026-03-21T06:50:09.551058+0000 mgr.x (mgr.14152) 110 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:11.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:10 vm04 bash[20194]: cluster 2026-03-21T06:50:09.551058+0000 mgr.x (mgr.14152) 110 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:10 vm02 bash[17657]: cluster 2026-03-21T06:50:09.551058+0000 mgr.x (mgr.14152) 110 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:10 vm02 bash[17657]: cluster 2026-03-21T06:50:09.551058+0000 mgr.x (mgr.14152) 110 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:10 vm07 bash[19945]: cluster 2026-03-21T06:50:09.551058+0000 mgr.x (mgr.14152) 110 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:10 vm07 bash[19945]: cluster 2026-03-21T06:50:09.551058+0000 mgr.x (mgr.14152) 110 : cluster [DBG] pgmap v59: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:13.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:12 vm04 bash[20194]: cluster 2026-03-21T06:50:11.551344+0000 mgr.x (mgr.14152) 111 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:13.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:12 vm04 bash[20194]: cluster 2026-03-21T06:50:11.551344+0000 mgr.x (mgr.14152) 111 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:12 vm02 bash[17657]: cluster 2026-03-21T06:50:11.551344+0000 mgr.x (mgr.14152) 111 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:12 vm02 bash[17657]: cluster 2026-03-21T06:50:11.551344+0000 mgr.x (mgr.14152) 111 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:12 vm07 bash[19945]: cluster 2026-03-21T06:50:11.551344+0000 mgr.x (mgr.14152) 111 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:12 vm07 bash[19945]: cluster 2026-03-21T06:50:11.551344+0000 mgr.x (mgr.14152) 111 : cluster [DBG] pgmap v60: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:15.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:14 vm04 bash[20194]: cluster 2026-03-21T06:50:13.551603+0000 mgr.x (mgr.14152) 112 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:15.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:14 vm04 bash[20194]: cluster 2026-03-21T06:50:13.551603+0000 mgr.x (mgr.14152) 112 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:15.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:14 vm02 bash[17657]: cluster 2026-03-21T06:50:13.551603+0000 mgr.x (mgr.14152) 112 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:15.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:14 vm02 bash[17657]: cluster 2026-03-21T06:50:13.551603+0000 mgr.x (mgr.14152) 112 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:14 vm07 bash[19945]: cluster 2026-03-21T06:50:13.551603+0000 mgr.x (mgr.14152) 112 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:14 vm07 bash[19945]: cluster 2026-03-21T06:50:13.551603+0000 mgr.x (mgr.14152) 112 : cluster [DBG] pgmap v61: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:17.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:16 vm04 bash[20194]: cluster 2026-03-21T06:50:15.551830+0000 mgr.x (mgr.14152) 113 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:17.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:16 vm04 bash[20194]: cluster 2026-03-21T06:50:15.551830+0000 mgr.x (mgr.14152) 113 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:17.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:16 vm02 bash[17657]: cluster 2026-03-21T06:50:15.551830+0000 mgr.x (mgr.14152) 113 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:17.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:16 vm02 bash[17657]: cluster 2026-03-21T06:50:15.551830+0000 mgr.x (mgr.14152) 113 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:16 vm07 bash[19945]: cluster 2026-03-21T06:50:15.551830+0000 mgr.x (mgr.14152) 113 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:16 vm07 bash[19945]: cluster 2026-03-21T06:50:15.551830+0000 mgr.x (mgr.14152) 113 : cluster [DBG] pgmap v62: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:18.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:17 vm04 bash[20194]: cluster 2026-03-21T06:50:17.552168+0000 mgr.x (mgr.14152) 114 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:18.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:17 vm04 bash[20194]: cluster 2026-03-21T06:50:17.552168+0000 mgr.x (mgr.14152) 114 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:18.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:17 vm02 bash[17657]: cluster 2026-03-21T06:50:17.552168+0000 mgr.x (mgr.14152) 114 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:18.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:17 vm02 bash[17657]: cluster 2026-03-21T06:50:17.552168+0000 mgr.x (mgr.14152) 114 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:18.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:17 vm07 bash[19945]: cluster 2026-03-21T06:50:17.552168+0000 mgr.x (mgr.14152) 114 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:18.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:17 vm07 bash[19945]: cluster 2026-03-21T06:50:17.552168+0000 mgr.x (mgr.14152) 114 : cluster [DBG] pgmap v63: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:20.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:20 vm02 bash[17657]: cluster 2026-03-21T06:50:19.552389+0000 mgr.x (mgr.14152) 115 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:20.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:20 vm02 bash[17657]: cluster 2026-03-21T06:50:19.552389+0000 mgr.x (mgr.14152) 115 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:20.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:20 vm07 bash[19945]: cluster 2026-03-21T06:50:19.552389+0000 mgr.x (mgr.14152) 115 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:20.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:20 vm07 bash[19945]: cluster 2026-03-21T06:50:19.552389+0000 mgr.x (mgr.14152) 115 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:21.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:20 vm04 bash[20194]: cluster 2026-03-21T06:50:19.552389+0000 mgr.x (mgr.14152) 115 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:21.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:20 vm04 bash[20194]: cluster 2026-03-21T06:50:19.552389+0000 mgr.x (mgr.14152) 115 : cluster [DBG] pgmap v64: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:22.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:22 vm02 bash[17657]: cluster 2026-03-21T06:50:21.552646+0000 mgr.x (mgr.14152) 116 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:22.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:22 vm02 bash[17657]: cluster 2026-03-21T06:50:21.552646+0000 mgr.x (mgr.14152) 116 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:22.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:22 vm07 bash[19945]: cluster 2026-03-21T06:50:21.552646+0000 mgr.x (mgr.14152) 116 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:22.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:22 vm07 bash[19945]: cluster 2026-03-21T06:50:21.552646+0000 mgr.x (mgr.14152) 116 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:23.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:22 vm04 bash[20194]: cluster 2026-03-21T06:50:21.552646+0000 mgr.x (mgr.14152) 116 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:23.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:22 vm04 bash[20194]: cluster 2026-03-21T06:50:21.552646+0000 mgr.x (mgr.14152) 116 : cluster [DBG] pgmap v65: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:24 vm02 bash[17657]: cluster 2026-03-21T06:50:23.552947+0000 mgr.x (mgr.14152) 117 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:24 vm02 bash[17657]: cluster 2026-03-21T06:50:23.552947+0000 mgr.x (mgr.14152) 117 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:24.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:24 vm07 bash[19945]: cluster 2026-03-21T06:50:23.552947+0000 mgr.x (mgr.14152) 117 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:24.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:24 vm07 bash[19945]: cluster 2026-03-21T06:50:23.552947+0000 mgr.x (mgr.14152) 117 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:25.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:24 vm04 bash[20194]: cluster 2026-03-21T06:50:23.552947+0000 mgr.x (mgr.14152) 117 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:25.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:24 vm04 bash[20194]: cluster 2026-03-21T06:50:23.552947+0000 mgr.x (mgr.14152) 117 : cluster [DBG] pgmap v66: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:26.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:26 vm02 bash[17657]: cluster 2026-03-21T06:50:25.553186+0000 mgr.x (mgr.14152) 118 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:26.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:26 vm02 bash[17657]: cluster 2026-03-21T06:50:25.553186+0000 mgr.x (mgr.14152) 118 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:26.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:26 vm07 bash[19945]: cluster 2026-03-21T06:50:25.553186+0000 mgr.x (mgr.14152) 118 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:26.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:26 vm07 bash[19945]: cluster 2026-03-21T06:50:25.553186+0000 mgr.x (mgr.14152) 118 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:27.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:26 vm04 bash[20194]: cluster 2026-03-21T06:50:25.553186+0000 mgr.x (mgr.14152) 118 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:27.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:26 vm04 bash[20194]: cluster 2026-03-21T06:50:25.553186+0000 mgr.x (mgr.14152) 118 : cluster [DBG] pgmap v67: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:28.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:28 vm02 bash[17657]: cluster 2026-03-21T06:50:27.553471+0000 mgr.x (mgr.14152) 119 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:28.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:28 vm02 bash[17657]: cluster 2026-03-21T06:50:27.553471+0000 mgr.x (mgr.14152) 119 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:28.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:28 vm07 bash[19945]: cluster 2026-03-21T06:50:27.553471+0000 mgr.x (mgr.14152) 119 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:28.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:28 vm07 bash[19945]: cluster 2026-03-21T06:50:27.553471+0000 mgr.x (mgr.14152) 119 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:29.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:28 vm04 bash[20194]: cluster 2026-03-21T06:50:27.553471+0000 mgr.x (mgr.14152) 119 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:29.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:28 vm04 bash[20194]: cluster 2026-03-21T06:50:27.553471+0000 mgr.x (mgr.14152) 119 : cluster [DBG] pgmap v68: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:30 vm02 bash[17657]: cluster 2026-03-21T06:50:29.553737+0000 mgr.x (mgr.14152) 120 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:30 vm02 bash[17657]: cluster 2026-03-21T06:50:29.553737+0000 mgr.x (mgr.14152) 120 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:30 vm07 bash[19945]: cluster 2026-03-21T06:50:29.553737+0000 mgr.x (mgr.14152) 120 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:30 vm07 bash[19945]: cluster 2026-03-21T06:50:29.553737+0000 mgr.x (mgr.14152) 120 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:31.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:30 vm04 bash[20194]: cluster 2026-03-21T06:50:29.553737+0000 mgr.x (mgr.14152) 120 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:30 vm04 bash[20194]: cluster 2026-03-21T06:50:29.553737+0000 mgr.x (mgr.14152) 120 : cluster [DBG] pgmap v69: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:33.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:32 vm04 bash[20194]: cluster 2026-03-21T06:50:31.553983+0000 mgr.x (mgr.14152) 121 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:33.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:32 vm04 bash[20194]: cluster 2026-03-21T06:50:31.553983+0000 mgr.x (mgr.14152) 121 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:33.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:32 vm02 bash[17657]: cluster 2026-03-21T06:50:31.553983+0000 mgr.x (mgr.14152) 121 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:33.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:32 vm02 bash[17657]: cluster 2026-03-21T06:50:31.553983+0000 mgr.x (mgr.14152) 121 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:32 vm07 bash[19945]: cluster 2026-03-21T06:50:31.553983+0000 mgr.x (mgr.14152) 121 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:32 vm07 bash[19945]: cluster 2026-03-21T06:50:31.553983+0000 mgr.x (mgr.14152) 121 : cluster [DBG] pgmap v70: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:35.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:34 vm04 bash[20194]: cluster 2026-03-21T06:50:33.554247+0000 mgr.x (mgr.14152) 122 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:35.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:34 vm04 bash[20194]: cluster 2026-03-21T06:50:33.554247+0000 mgr.x (mgr.14152) 122 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:35.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:34 vm02 bash[17657]: cluster 2026-03-21T06:50:33.554247+0000 mgr.x (mgr.14152) 122 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:35.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:34 vm02 bash[17657]: cluster 2026-03-21T06:50:33.554247+0000 mgr.x (mgr.14152) 122 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:35.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:34 vm07 bash[19945]: cluster 2026-03-21T06:50:33.554247+0000 mgr.x (mgr.14152) 122 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:35.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:34 vm07 bash[19945]: cluster 2026-03-21T06:50:33.554247+0000 mgr.x (mgr.14152) 122 : cluster [DBG] pgmap v71: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:37.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:36 vm04 bash[20194]: cluster 2026-03-21T06:50:35.554463+0000 mgr.x (mgr.14152) 123 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:37.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:36 vm04 bash[20194]: cluster 2026-03-21T06:50:35.554463+0000 mgr.x (mgr.14152) 123 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:37.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:36 vm02 bash[17657]: cluster 2026-03-21T06:50:35.554463+0000 mgr.x (mgr.14152) 123 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:37.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:36 vm02 bash[17657]: cluster 2026-03-21T06:50:35.554463+0000 mgr.x (mgr.14152) 123 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:36 vm07 bash[19945]: cluster 2026-03-21T06:50:35.554463+0000 mgr.x (mgr.14152) 123 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:36 vm07 bash[19945]: cluster 2026-03-21T06:50:35.554463+0000 mgr.x (mgr.14152) 123 : cluster [DBG] pgmap v72: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:39.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:38 vm04 bash[20194]: cluster 2026-03-21T06:50:37.554676+0000 mgr.x (mgr.14152) 124 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:39.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:38 vm04 bash[20194]: cluster 2026-03-21T06:50:37.554676+0000 mgr.x (mgr.14152) 124 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:39.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:38 vm02 bash[17657]: cluster 2026-03-21T06:50:37.554676+0000 mgr.x (mgr.14152) 124 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:39.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:38 vm02 bash[17657]: cluster 2026-03-21T06:50:37.554676+0000 mgr.x (mgr.14152) 124 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:38 vm07 bash[19945]: cluster 2026-03-21T06:50:37.554676+0000 mgr.x (mgr.14152) 124 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:38 vm07 bash[19945]: cluster 2026-03-21T06:50:37.554676+0000 mgr.x (mgr.14152) 124 : cluster [DBG] pgmap v73: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:41.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:40 vm04 bash[20194]: cluster 2026-03-21T06:50:39.554861+0000 mgr.x (mgr.14152) 125 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:41.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:40 vm04 bash[20194]: cluster 2026-03-21T06:50:39.554861+0000 mgr.x (mgr.14152) 125 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:41.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:40 vm02 bash[17657]: cluster 2026-03-21T06:50:39.554861+0000 mgr.x (mgr.14152) 125 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:41.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:40 vm02 bash[17657]: cluster 2026-03-21T06:50:39.554861+0000 mgr.x (mgr.14152) 125 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:40 vm07 bash[19945]: cluster 2026-03-21T06:50:39.554861+0000 mgr.x (mgr.14152) 125 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:40 vm07 bash[19945]: cluster 2026-03-21T06:50:39.554861+0000 mgr.x (mgr.14152) 125 : cluster [DBG] pgmap v74: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:43.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:42 vm04 bash[20194]: cluster 2026-03-21T06:50:41.555044+0000 mgr.x (mgr.14152) 126 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:42 vm04 bash[20194]: cluster 2026-03-21T06:50:41.555044+0000 mgr.x (mgr.14152) 126 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:43.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:42 vm02 bash[17657]: cluster 2026-03-21T06:50:41.555044+0000 mgr.x (mgr.14152) 126 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:43.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:42 vm02 bash[17657]: cluster 2026-03-21T06:50:41.555044+0000 mgr.x (mgr.14152) 126 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:42 vm07 bash[19945]: cluster 2026-03-21T06:50:41.555044+0000 mgr.x (mgr.14152) 126 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:42 vm07 bash[19945]: cluster 2026-03-21T06:50:41.555044+0000 mgr.x (mgr.14152) 126 : cluster [DBG] pgmap v75: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:45.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:44 vm04 bash[20194]: cluster 2026-03-21T06:50:43.555273+0000 mgr.x (mgr.14152) 127 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:45.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:44 vm04 bash[20194]: cluster 2026-03-21T06:50:43.555273+0000 mgr.x (mgr.14152) 127 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:45.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:44 vm02 bash[17657]: cluster 2026-03-21T06:50:43.555273+0000 mgr.x (mgr.14152) 127 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:45.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:44 vm02 bash[17657]: cluster 2026-03-21T06:50:43.555273+0000 mgr.x (mgr.14152) 127 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:45.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:44 vm07 bash[19945]: cluster 2026-03-21T06:50:43.555273+0000 mgr.x (mgr.14152) 127 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:45.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:44 vm07 bash[19945]: cluster 2026-03-21T06:50:43.555273+0000 mgr.x (mgr.14152) 127 : cluster [DBG] pgmap v76: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:46.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:45 vm04 bash[20194]: cluster 2026-03-21T06:50:45.555472+0000 mgr.x (mgr.14152) 128 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:46.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:45 vm04 bash[20194]: cluster 2026-03-21T06:50:45.555472+0000 mgr.x (mgr.14152) 128 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:46.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:45 vm02 bash[17657]: cluster 2026-03-21T06:50:45.555472+0000 mgr.x (mgr.14152) 128 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:46.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:45 vm02 bash[17657]: cluster 2026-03-21T06:50:45.555472+0000 mgr.x (mgr.14152) 128 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:46.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:45 vm07 bash[19945]: cluster 2026-03-21T06:50:45.555472+0000 mgr.x (mgr.14152) 128 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:46.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:45 vm07 bash[19945]: cluster 2026-03-21T06:50:45.555472+0000 mgr.x (mgr.14152) 128 : cluster [DBG] pgmap v77: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:47.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:47 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:50:47.646 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:50:47 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:50:48.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:48 vm02 bash[17657]: cluster 2026-03-21T06:50:47.555679+0000 mgr.x (mgr.14152) 129 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:48.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:48 vm02 bash[17657]: cluster 2026-03-21T06:50:47.555679+0000 mgr.x (mgr.14152) 129 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:48.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:48 vm02 bash[17657]: audit 2026-03-21T06:50:47.607910+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:48.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:48 vm02 bash[17657]: audit 2026-03-21T06:50:47.607910+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:48.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:48 vm02 bash[17657]: audit 2026-03-21T06:50:47.616116+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:48.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:48 vm02 bash[17657]: audit 2026-03-21T06:50:47.616116+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:48 vm07 bash[19945]: cluster 2026-03-21T06:50:47.555679+0000 mgr.x (mgr.14152) 129 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:48 vm07 bash[19945]: cluster 2026-03-21T06:50:47.555679+0000 mgr.x (mgr.14152) 129 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:48 vm07 bash[19945]: audit 2026-03-21T06:50:47.607910+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:48 vm07 bash[19945]: audit 2026-03-21T06:50:47.607910+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:48 vm07 bash[19945]: audit 2026-03-21T06:50:47.616116+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:48 vm07 bash[19945]: audit 2026-03-21T06:50:47.616116+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:49.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:48 vm04 bash[20194]: cluster 2026-03-21T06:50:47.555679+0000 mgr.x (mgr.14152) 129 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:49.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:48 vm04 bash[20194]: cluster 2026-03-21T06:50:47.555679+0000 mgr.x (mgr.14152) 129 : cluster [DBG] pgmap v78: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:49.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:48 vm04 bash[20194]: audit 2026-03-21T06:50:47.607910+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:49.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:48 vm04 bash[20194]: audit 2026-03-21T06:50:47.607910+0000 mon.a (mon.0) 301 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:49.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:48 vm04 bash[20194]: audit 2026-03-21T06:50:47.616116+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:49.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:48 vm04 bash[20194]: audit 2026-03-21T06:50:47.616116+0000 mon.a (mon.0) 302 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:49.082 INFO:teuthology.orchestra.run.vm02.stdout:Created osd(s) 0 on host 'vm02' 2026-03-21T06:50:49.168 DEBUG:teuthology.orchestra.run.vm02:osd.0> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.0.service 2026-03-21T06:50:49.168 INFO:tasks.cephadm:Deploying osd.1 on vm02 with /dev/vg_nvme/lv_3... 2026-03-21T06:50:49.168 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- lvm zap /dev/vg_nvme/lv_3 2026-03-21T06:50:49.288 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:49 vm02 bash[29218]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 2026-03-21T06:50:49.289 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:49 vm02 bash[29218]: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/vg_nvme/lv_4 --path /var/lib/ceph/osd/ceph-0 --no-mon-config 2026-03-21T06:50:49.289 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:49 vm02 bash[29218]: Running command: /usr/bin/ln -snf /dev/vg_nvme/lv_4 /var/lib/ceph/osd/ceph-0/block 2026-03-21T06:50:49.479 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:50:49.506 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:49 vm02 bash[29218]: Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-0/block 2026-03-21T06:50:49.506 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:49 vm02 bash[29218]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-3 2026-03-21T06:50:49.506 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:49 vm02 bash[29218]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-0 2026-03-21T06:50:49.506 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:49 vm02 bash[29218]: --> ceph-volume lvm activate successful for osd ID: 0 2026-03-21T06:50:49.896 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:49 vm02 bash[30128]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-0 2026-03-21T06:50:49.896 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:49 vm02 bash[30274]: debug 2026-03-21T06:50:49.694+0000 7f8aa5695640 1 -- 192.168.123.102:0/1917300771 <== mon.0 v2:192.168.123.102:3300/0 4 ==== auth_reply(proto 2 0 (0) Success) ==== 194+0+0 (secure 0 0 0) 0x559e10f2b860 con 0x559e0ce61c00 2026-03-21T06:50:50.306 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:50 vm02 bash[17657]: audit 2026-03-21T06:50:49.065910+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.306 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:50 vm02 bash[17657]: audit 2026-03-21T06:50:49.065910+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.306 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:50 vm02 bash[17657]: audit 2026-03-21T06:50:49.073750+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.306 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:50 vm02 bash[17657]: audit 2026-03-21T06:50:49.073750+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.306 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:50 vm02 bash[17657]: cluster 2026-03-21T06:50:49.556130+0000 mgr.x (mgr.14152) 130 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:50.306 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:50 vm02 bash[17657]: cluster 2026-03-21T06:50:49.556130+0000 mgr.x (mgr.14152) 130 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:50.306 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:50 vm02 bash[30274]: debug 2026-03-21T06:50:50.038+0000 7f8aa7eff8c0 -1 Falling back to public interface 2026-03-21T06:50:50.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:50 vm07 bash[19945]: audit 2026-03-21T06:50:49.065910+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:50 vm07 bash[19945]: audit 2026-03-21T06:50:49.065910+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:50 vm07 bash[19945]: audit 2026-03-21T06:50:49.073750+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:50 vm07 bash[19945]: audit 2026-03-21T06:50:49.073750+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:50 vm07 bash[19945]: cluster 2026-03-21T06:50:49.556130+0000 mgr.x (mgr.14152) 130 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:50.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:50 vm07 bash[19945]: cluster 2026-03-21T06:50:49.556130+0000 mgr.x (mgr.14152) 130 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:50.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:50 vm04 bash[20194]: audit 2026-03-21T06:50:49.065910+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:50 vm04 bash[20194]: audit 2026-03-21T06:50:49.065910+0000 mon.a (mon.0) 303 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:50 vm04 bash[20194]: audit 2026-03-21T06:50:49.073750+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:50 vm04 bash[20194]: audit 2026-03-21T06:50:49.073750+0000 mon.a (mon.0) 304 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:50.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:50 vm04 bash[20194]: cluster 2026-03-21T06:50:49.556130+0000 mgr.x (mgr.14152) 130 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:50.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:50 vm04 bash[20194]: cluster 2026-03-21T06:50:49.556130+0000 mgr.x (mgr.14152) 130 : cluster [DBG] pgmap v79: 0 pgs: ; 0 B data, 26 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:50.646 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:50 vm02 bash[30274]: debug 2026-03-21T06:50:50.326+0000 7f8aa7eff8c0 -1 osd.0 9 log_to_monitors true 2026-03-21T06:50:50.795 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:50:50.812 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch daemon add osd vm02:vg_nvme/lv_3 --skip-validation 2026-03-21T06:50:51.047 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:50:51.083 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:51 vm02 bash[17657]: audit 2026-03-21T06:50:50.332670+0000 mon.a (mon.0) 305 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:50:51.083 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:51 vm02 bash[17657]: audit 2026-03-21T06:50:50.332670+0000 mon.a (mon.0) 305 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:50:51.083 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:51 vm02 bash[17657]: audit 2026-03-21T06:50:50.368044+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.083 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:51 vm02 bash[17657]: audit 2026-03-21T06:50:50.368044+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.083 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:51 vm02 bash[17657]: audit 2026-03-21T06:50:50.377938+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.083 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:51 vm02 bash[17657]: audit 2026-03-21T06:50:50.377938+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.083 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:51 vm02 bash[17657]: audit 2026-03-21T06:50:50.396242+0000 mon.a (mon.0) 308 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:50:51.083 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:51 vm02 bash[17657]: audit 2026-03-21T06:50:50.396242+0000 mon.a (mon.0) 308 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:50:51.396 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:51 vm02 bash[30274]: debug 2026-03-21T06:50:51.102+0000 7f8a9ec66640 -1 osd.0 9 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-21T06:50:51.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:51 vm07 bash[19945]: audit 2026-03-21T06:50:50.332670+0000 mon.a (mon.0) 305 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:50:51.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:51 vm07 bash[19945]: audit 2026-03-21T06:50:50.332670+0000 mon.a (mon.0) 305 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:50:51.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:51 vm07 bash[19945]: audit 2026-03-21T06:50:50.368044+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:51 vm07 bash[19945]: audit 2026-03-21T06:50:50.368044+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:51 vm07 bash[19945]: audit 2026-03-21T06:50:50.377938+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:51 vm07 bash[19945]: audit 2026-03-21T06:50:50.377938+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:51 vm07 bash[19945]: audit 2026-03-21T06:50:50.396242+0000 mon.a (mon.0) 308 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:50:51.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:51 vm07 bash[19945]: audit 2026-03-21T06:50:50.396242+0000 mon.a (mon.0) 308 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:50:51.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:51 vm04 bash[20194]: audit 2026-03-21T06:50:50.332670+0000 mon.a (mon.0) 305 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:50:51.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:51 vm04 bash[20194]: audit 2026-03-21T06:50:50.332670+0000 mon.a (mon.0) 305 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]} : dispatch 2026-03-21T06:50:51.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:51 vm04 bash[20194]: audit 2026-03-21T06:50:50.368044+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:51 vm04 bash[20194]: audit 2026-03-21T06:50:50.368044+0000 mon.a (mon.0) 306 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:51 vm04 bash[20194]: audit 2026-03-21T06:50:50.377938+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:51 vm04 bash[20194]: audit 2026-03-21T06:50:50.377938+0000 mon.a (mon.0) 307 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:51.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:51 vm04 bash[20194]: audit 2026-03-21T06:50:50.396242+0000 mon.a (mon.0) 308 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:50:51.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:51 vm04 bash[20194]: audit 2026-03-21T06:50:50.396242+0000 mon.a (mon.0) 308 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.082010+0000 mon.a (mon.0) 309 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.082010+0000 mon.a (mon.0) 309 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cluster 2026-03-21T06:50:51.084238+0000 mon.a (mon.0) 310 : cluster [DBG] osdmap e10: 1 total, 1 up, 1 in 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cluster 2026-03-21T06:50:51.084238+0000 mon.a (mon.0) 310 : cluster [DBG] osdmap e10: 1 total, 1 up, 1 in 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.084384+0000 mon.a (mon.0) 311 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.084384+0000 mon.a (mon.0) 311 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.360012+0000 mgr.x (mgr.14152) 131 : audit [DBG] from='client.14259 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.360012+0000 mgr.x (mgr.14152) 131 : audit [DBG] from='client.14259 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cephadm 2026-03-21T06:50:51.360786+0000 mgr.x (mgr.14152) 132 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cephadm 2026-03-21T06:50:51.360786+0000 mgr.x (mgr.14152) 132 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cephadm 2026-03-21T06:50:51.360852+0000 mgr.x (mgr.14152) 133 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_3'] 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cephadm 2026-03-21T06:50:51.360852+0000 mgr.x (mgr.14152) 133 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_3'] 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.365053+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.365053+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cephadm 2026-03-21T06:50:51.365401+0000 mgr.x (mgr.14152) 134 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cephadm 2026-03-21T06:50:51.365401+0000 mgr.x (mgr.14152) 134 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cephadm 2026-03-21T06:50:51.365509+0000 mgr.x (mgr.14152) 135 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cephadm 2026-03-21T06:50:51.365509+0000 mgr.x (mgr.14152) 135 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.368605+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.368605+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.369410+0000 mon.a (mon.0) 314 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:52.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.369410+0000 mon.a (mon.0) 314 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:52.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.370702+0000 mon.a (mon.0) 315 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:52.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.370702+0000 mon.a (mon.0) 315 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:52.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.371139+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:52.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: audit 2026-03-21T06:50:51.371139+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:52.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cluster 2026-03-21T06:50:51.556518+0000 mgr.x (mgr.14152) 136 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:52.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:52 vm02 bash[17657]: cluster 2026-03-21T06:50:51.556518+0000 mgr.x (mgr.14152) 136 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.082010+0000 mon.a (mon.0) 309 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.082010+0000 mon.a (mon.0) 309 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cluster 2026-03-21T06:50:51.084238+0000 mon.a (mon.0) 310 : cluster [DBG] osdmap e10: 1 total, 1 up, 1 in 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cluster 2026-03-21T06:50:51.084238+0000 mon.a (mon.0) 310 : cluster [DBG] osdmap e10: 1 total, 1 up, 1 in 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.084384+0000 mon.a (mon.0) 311 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.084384+0000 mon.a (mon.0) 311 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.360012+0000 mgr.x (mgr.14152) 131 : audit [DBG] from='client.14259 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.360012+0000 mgr.x (mgr.14152) 131 : audit [DBG] from='client.14259 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cephadm 2026-03-21T06:50:51.360786+0000 mgr.x (mgr.14152) 132 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cephadm 2026-03-21T06:50:51.360786+0000 mgr.x (mgr.14152) 132 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cephadm 2026-03-21T06:50:51.360852+0000 mgr.x (mgr.14152) 133 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_3'] 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cephadm 2026-03-21T06:50:51.360852+0000 mgr.x (mgr.14152) 133 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_3'] 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.365053+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.365053+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cephadm 2026-03-21T06:50:51.365401+0000 mgr.x (mgr.14152) 134 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cephadm 2026-03-21T06:50:51.365401+0000 mgr.x (mgr.14152) 134 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cephadm 2026-03-21T06:50:51.365509+0000 mgr.x (mgr.14152) 135 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cephadm 2026-03-21T06:50:51.365509+0000 mgr.x (mgr.14152) 135 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.368605+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.368605+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.369410+0000 mon.a (mon.0) 314 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.369410+0000 mon.a (mon.0) 314 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.370702+0000 mon.a (mon.0) 315 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.370702+0000 mon.a (mon.0) 315 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.371139+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: audit 2026-03-21T06:50:51.371139+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cluster 2026-03-21T06:50:51.556518+0000 mgr.x (mgr.14152) 136 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:52.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:52 vm07 bash[19945]: cluster 2026-03-21T06:50:51.556518+0000 mgr.x (mgr.14152) 136 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.082010+0000 mon.a (mon.0) 309 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.082010+0000 mon.a (mon.0) 309 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["0"]}]': finished 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cluster 2026-03-21T06:50:51.084238+0000 mon.a (mon.0) 310 : cluster [DBG] osdmap e10: 1 total, 1 up, 1 in 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cluster 2026-03-21T06:50:51.084238+0000 mon.a (mon.0) 310 : cluster [DBG] osdmap e10: 1 total, 1 up, 1 in 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.084384+0000 mon.a (mon.0) 311 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.084384+0000 mon.a (mon.0) 311 : audit [INF] from='osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539]' entity='osd.0' cmd={"prefix": "osd crush create-or-move", "id": 0, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.360012+0000 mgr.x (mgr.14152) 131 : audit [DBG] from='client.14259 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.360012+0000 mgr.x (mgr.14152) 131 : audit [DBG] from='client.14259 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm02:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cephadm 2026-03-21T06:50:51.360786+0000 mgr.x (mgr.14152) 132 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cephadm 2026-03-21T06:50:51.360786+0000 mgr.x (mgr.14152) 132 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cephadm 2026-03-21T06:50:51.360852+0000 mgr.x (mgr.14152) 133 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_3'] 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cephadm 2026-03-21T06:50:51.360852+0000 mgr.x (mgr.14152) 133 : cephadm [INF] Creating OSDs with service ID: default on vm02:['vg_nvme/lv_3'] 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.365053+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.365053+0000 mon.a (mon.0) 312 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cephadm 2026-03-21T06:50:51.365401+0000 mgr.x (mgr.14152) 134 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cephadm 2026-03-21T06:50:51.365401+0000 mgr.x (mgr.14152) 134 : cephadm [INF] Marking host: vm02 for OSDSpec preview refresh. 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cephadm 2026-03-21T06:50:51.365509+0000 mgr.x (mgr.14152) 135 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cephadm 2026-03-21T06:50:51.365509+0000 mgr.x (mgr.14152) 135 : cephadm [INF] Saving service osd.default spec with placement vm02 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.368605+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.368605+0000 mon.a (mon.0) 313 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.369410+0000 mon.a (mon.0) 314 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.369410+0000 mon.a (mon.0) 314 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.370702+0000 mon.a (mon.0) 315 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.370702+0000 mon.a (mon.0) 315 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.371139+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:52.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: audit 2026-03-21T06:50:51.371139+0000 mon.a (mon.0) 316 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:52.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cluster 2026-03-21T06:50:51.556518+0000 mgr.x (mgr.14152) 136 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:52.505 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:52 vm04 bash[20194]: cluster 2026-03-21T06:50:51.556518+0000 mgr.x (mgr.14152) 136 : cluster [DBG] pgmap v81: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.083163+0000 mon.a (mon.0) 317 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.083163+0000 mon.a (mon.0) 317 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.083195+0000 mon.a (mon.0) 318 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.083195+0000 mon.a (mon.0) 318 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.083206+0000 mon.a (mon.0) 319 : cluster [WRN] Health check failed: 1 root (1 osds) down (OSD_ROOT_DOWN) 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.083206+0000 mon.a (mon.0) 319 : cluster [WRN] Health check failed: 1 root (1 osds) down (OSD_ROOT_DOWN) 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.094943+0000 mon.a (mon.0) 320 : cluster [DBG] osdmap e11: 1 total, 0 up, 1 in 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.094943+0000 mon.a (mon.0) 320 : cluster [DBG] osdmap e11: 1 total, 0 up, 1 in 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.225308+0000 mon.a (mon.0) 321 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.225308+0000 mon.a (mon.0) 321 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.225833+0000 mon.a (mon.0) 322 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.225833+0000 mon.a (mon.0) 322 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:50:53.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.225847+0000 mon.a (mon.0) 323 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.225847+0000 mon.a (mon.0) 323 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.225854+0000 mon.a (mon.0) 324 : cluster [INF] Health check cleared: OSD_ROOT_DOWN (was: 1 root (1 osds) down) 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.225854+0000 mon.a (mon.0) 324 : cluster [INF] Health check cleared: OSD_ROOT_DOWN (was: 1 root (1 osds) down) 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.225860+0000 mon.a (mon.0) 325 : cluster [INF] Cluster is now healthy 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.225860+0000 mon.a (mon.0) 325 : cluster [INF] Cluster is now healthy 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.228516+0000 mon.a (mon.0) 326 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"}]': finished 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.228516+0000 mon.a (mon.0) 326 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"}]': finished 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.230118+0000 mon.c (mon.1) 4 : audit [INF] from='client.? 192.168.123.102:0/1629133887' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.230118+0000 mon.c (mon.1) 4 : audit [INF] from='client.? 192.168.123.102:0/1629133887' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.231559+0000 mon.a (mon.0) 327 : cluster [INF] osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539] boot 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.231559+0000 mon.a (mon.0) 327 : cluster [INF] osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539] boot 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.231617+0000 mon.a (mon.0) 328 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: cluster 2026-03-21T06:50:52.231617+0000 mon.a (mon.0) 328 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.231748+0000 mon.a (mon.0) 329 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.231748+0000 mon.a (mon.0) 329 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.231851+0000 mon.a (mon.0) 330 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.231851+0000 mon.a (mon.0) 330 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.599316+0000 mon.c (mon.1) 5 : audit [DBG] from='client.? 192.168.123.102:0/3273465308' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:50:53.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:53 vm02 bash[17657]: audit 2026-03-21T06:50:52.599316+0000 mon.c (mon.1) 5 : audit [DBG] from='client.? 192.168.123.102:0/3273465308' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:50:53.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.083163+0000 mon.a (mon.0) 317 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.083163+0000 mon.a (mon.0) 317 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.083195+0000 mon.a (mon.0) 318 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.083195+0000 mon.a (mon.0) 318 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.083206+0000 mon.a (mon.0) 319 : cluster [WRN] Health check failed: 1 root (1 osds) down (OSD_ROOT_DOWN) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.083206+0000 mon.a (mon.0) 319 : cluster [WRN] Health check failed: 1 root (1 osds) down (OSD_ROOT_DOWN) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.094943+0000 mon.a (mon.0) 320 : cluster [DBG] osdmap e11: 1 total, 0 up, 1 in 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.094943+0000 mon.a (mon.0) 320 : cluster [DBG] osdmap e11: 1 total, 0 up, 1 in 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.225308+0000 mon.a (mon.0) 321 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.225308+0000 mon.a (mon.0) 321 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.225833+0000 mon.a (mon.0) 322 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.225833+0000 mon.a (mon.0) 322 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.225847+0000 mon.a (mon.0) 323 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.225847+0000 mon.a (mon.0) 323 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.225854+0000 mon.a (mon.0) 324 : cluster [INF] Health check cleared: OSD_ROOT_DOWN (was: 1 root (1 osds) down) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.225854+0000 mon.a (mon.0) 324 : cluster [INF] Health check cleared: OSD_ROOT_DOWN (was: 1 root (1 osds) down) 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.225860+0000 mon.a (mon.0) 325 : cluster [INF] Cluster is now healthy 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.225860+0000 mon.a (mon.0) 325 : cluster [INF] Cluster is now healthy 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.228516+0000 mon.a (mon.0) 326 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"}]': finished 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.228516+0000 mon.a (mon.0) 326 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"}]': finished 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.230118+0000 mon.c (mon.1) 4 : audit [INF] from='client.? 192.168.123.102:0/1629133887' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.230118+0000 mon.c (mon.1) 4 : audit [INF] from='client.? 192.168.123.102:0/1629133887' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.231559+0000 mon.a (mon.0) 327 : cluster [INF] osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539] boot 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.231559+0000 mon.a (mon.0) 327 : cluster [INF] osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539] boot 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.231617+0000 mon.a (mon.0) 328 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: cluster 2026-03-21T06:50:52.231617+0000 mon.a (mon.0) 328 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.231748+0000 mon.a (mon.0) 329 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.231748+0000 mon.a (mon.0) 329 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.231851+0000 mon.a (mon.0) 330 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.231851+0000 mon.a (mon.0) 330 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.599316+0000 mon.c (mon.1) 5 : audit [DBG] from='client.? 192.168.123.102:0/3273465308' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:50:53.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:53 vm07 bash[19945]: audit 2026-03-21T06:50:52.599316+0000 mon.c (mon.1) 5 : audit [DBG] from='client.? 192.168.123.102:0/3273465308' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.083163+0000 mon.a (mon.0) 317 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.083163+0000 mon.a (mon.0) 317 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.083195+0000 mon.a (mon.0) 318 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.083195+0000 mon.a (mon.0) 318 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.083206+0000 mon.a (mon.0) 319 : cluster [WRN] Health check failed: 1 root (1 osds) down (OSD_ROOT_DOWN) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.083206+0000 mon.a (mon.0) 319 : cluster [WRN] Health check failed: 1 root (1 osds) down (OSD_ROOT_DOWN) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.094943+0000 mon.a (mon.0) 320 : cluster [DBG] osdmap e11: 1 total, 0 up, 1 in 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.094943+0000 mon.a (mon.0) 320 : cluster [DBG] osdmap e11: 1 total, 0 up, 1 in 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.225308+0000 mon.a (mon.0) 321 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.225308+0000 mon.a (mon.0) 321 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.225833+0000 mon.a (mon.0) 322 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.225833+0000 mon.a (mon.0) 322 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.225847+0000 mon.a (mon.0) 323 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.225847+0000 mon.a (mon.0) 323 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.225854+0000 mon.a (mon.0) 324 : cluster [INF] Health check cleared: OSD_ROOT_DOWN (was: 1 root (1 osds) down) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.225854+0000 mon.a (mon.0) 324 : cluster [INF] Health check cleared: OSD_ROOT_DOWN (was: 1 root (1 osds) down) 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.225860+0000 mon.a (mon.0) 325 : cluster [INF] Cluster is now healthy 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.225860+0000 mon.a (mon.0) 325 : cluster [INF] Cluster is now healthy 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.228516+0000 mon.a (mon.0) 326 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"}]': finished 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.228516+0000 mon.a (mon.0) 326 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"}]': finished 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.230118+0000 mon.c (mon.1) 4 : audit [INF] from='client.? 192.168.123.102:0/1629133887' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.230118+0000 mon.c (mon.1) 4 : audit [INF] from='client.? 192.168.123.102:0/1629133887' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "38235c66-ddf2-470c-8f4a-dac1efac5f75"} : dispatch 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.231559+0000 mon.a (mon.0) 327 : cluster [INF] osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539] boot 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.231559+0000 mon.a (mon.0) 327 : cluster [INF] osd.0 [v2:192.168.123.102:6802/2802530539,v1:192.168.123.102:6803/2802530539] boot 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.231617+0000 mon.a (mon.0) 328 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: cluster 2026-03-21T06:50:52.231617+0000 mon.a (mon.0) 328 : cluster [DBG] osdmap e12: 2 total, 1 up, 2 in 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.231748+0000 mon.a (mon.0) 329 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.231748+0000 mon.a (mon.0) 329 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 0} : dispatch 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.231851+0000 mon.a (mon.0) 330 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.231851+0000 mon.a (mon.0) 330 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.599316+0000 mon.c (mon.1) 5 : audit [DBG] from='client.? 192.168.123.102:0/3273465308' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:50:53.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:53 vm04 bash[20194]: audit 2026-03-21T06:50:52.599316+0000 mon.c (mon.1) 5 : audit [DBG] from='client.? 192.168.123.102:0/3273465308' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:50:54.259 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:54 vm02 bash[17657]: cluster 2026-03-21T06:50:53.556785+0000 mgr.x (mgr.14152) 137 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:54.259 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:54 vm02 bash[17657]: cluster 2026-03-21T06:50:53.556785+0000 mgr.x (mgr.14152) 137 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:54.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:54 vm07 bash[19945]: cluster 2026-03-21T06:50:53.556785+0000 mgr.x (mgr.14152) 137 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:54.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:54 vm07 bash[19945]: cluster 2026-03-21T06:50:53.556785+0000 mgr.x (mgr.14152) 137 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:54.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:54 vm04 bash[20194]: cluster 2026-03-21T06:50:53.556785+0000 mgr.x (mgr.14152) 137 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:54.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:54 vm04 bash[20194]: cluster 2026-03-21T06:50:53.556785+0000 mgr.x (mgr.14152) 137 : cluster [DBG] pgmap v84: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:55.079 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:55 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:50:55.079 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:50:55 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:50:55.079 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:55 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:50:55.335 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:50:55 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:50:55.335 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:55 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:50:55.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:55 vm02 bash[17657]: audit 2026-03-21T06:50:54.299360+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:50:55.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:55 vm02 bash[17657]: audit 2026-03-21T06:50:54.299360+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:50:55.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:55 vm02 bash[17657]: audit 2026-03-21T06:50:54.299919+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:55.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:55 vm02 bash[17657]: audit 2026-03-21T06:50:54.299919+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:55.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:55 vm02 bash[17657]: cephadm 2026-03-21T06:50:54.300302+0000 mgr.x (mgr.14152) 138 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:50:55.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:55 vm02 bash[17657]: cephadm 2026-03-21T06:50:54.300302+0000 mgr.x (mgr.14152) 138 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:50:55.335 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:55 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:50:55.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:55 vm07 bash[19945]: audit 2026-03-21T06:50:54.299360+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:50:55.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:55 vm07 bash[19945]: audit 2026-03-21T06:50:54.299360+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:50:55.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:55 vm07 bash[19945]: audit 2026-03-21T06:50:54.299919+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:55.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:55 vm07 bash[19945]: audit 2026-03-21T06:50:54.299919+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:55.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:55 vm07 bash[19945]: cephadm 2026-03-21T06:50:54.300302+0000 mgr.x (mgr.14152) 138 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:50:55.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:55 vm07 bash[19945]: cephadm 2026-03-21T06:50:54.300302+0000 mgr.x (mgr.14152) 138 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:50:55.503 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:55 vm04 bash[20194]: audit 2026-03-21T06:50:54.299360+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:50:55.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:55 vm04 bash[20194]: audit 2026-03-21T06:50:54.299360+0000 mon.a (mon.0) 331 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:50:55.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:55 vm04 bash[20194]: audit 2026-03-21T06:50:54.299919+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:55.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:55 vm04 bash[20194]: audit 2026-03-21T06:50:54.299919+0000 mon.a (mon.0) 332 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:55.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:55 vm04 bash[20194]: cephadm 2026-03-21T06:50:54.300302+0000 mgr.x (mgr.14152) 138 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:50:55.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:55 vm04 bash[20194]: cephadm 2026-03-21T06:50:54.300302+0000 mgr.x (mgr.14152) 138 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:50:56.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:56 vm02 bash[17657]: audit 2026-03-21T06:50:55.285601+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:56 vm02 bash[17657]: audit 2026-03-21T06:50:55.285601+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:56 vm02 bash[17657]: audit 2026-03-21T06:50:55.290248+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:56 vm02 bash[17657]: audit 2026-03-21T06:50:55.290248+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:56 vm02 bash[17657]: audit 2026-03-21T06:50:55.427610+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:56 vm02 bash[17657]: audit 2026-03-21T06:50:55.427610+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:56 vm02 bash[17657]: audit 2026-03-21T06:50:55.434651+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:56 vm02 bash[17657]: audit 2026-03-21T06:50:55.434651+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:56 vm02 bash[17657]: cluster 2026-03-21T06:50:55.557012+0000 mgr.x (mgr.14152) 139 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:56.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:56 vm02 bash[17657]: cluster 2026-03-21T06:50:55.557012+0000 mgr.x (mgr.14152) 139 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:56.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:56 vm07 bash[19945]: audit 2026-03-21T06:50:55.285601+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:56 vm07 bash[19945]: audit 2026-03-21T06:50:55.285601+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:56 vm07 bash[19945]: audit 2026-03-21T06:50:55.290248+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:56 vm07 bash[19945]: audit 2026-03-21T06:50:55.290248+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:56 vm07 bash[19945]: audit 2026-03-21T06:50:55.427610+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:56 vm07 bash[19945]: audit 2026-03-21T06:50:55.427610+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:56 vm07 bash[19945]: audit 2026-03-21T06:50:55.434651+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:56 vm07 bash[19945]: audit 2026-03-21T06:50:55.434651+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:56 vm07 bash[19945]: cluster 2026-03-21T06:50:55.557012+0000 mgr.x (mgr.14152) 139 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:56.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:56 vm07 bash[19945]: cluster 2026-03-21T06:50:55.557012+0000 mgr.x (mgr.14152) 139 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:56.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:56 vm04 bash[20194]: audit 2026-03-21T06:50:55.285601+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:56 vm04 bash[20194]: audit 2026-03-21T06:50:55.285601+0000 mon.a (mon.0) 333 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:56 vm04 bash[20194]: audit 2026-03-21T06:50:55.290248+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:56 vm04 bash[20194]: audit 2026-03-21T06:50:55.290248+0000 mon.a (mon.0) 334 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:56 vm04 bash[20194]: audit 2026-03-21T06:50:55.427610+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:56 vm04 bash[20194]: audit 2026-03-21T06:50:55.427610+0000 mon.a (mon.0) 335 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:56 vm04 bash[20194]: audit 2026-03-21T06:50:55.434651+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:56 vm04 bash[20194]: audit 2026-03-21T06:50:55.434651+0000 mon.a (mon.0) 336 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:56.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:56 vm04 bash[20194]: cluster 2026-03-21T06:50:55.557012+0000 mgr.x (mgr.14152) 139 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:56.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:56 vm04 bash[20194]: cluster 2026-03-21T06:50:55.557012+0000 mgr.x (mgr.14152) 139 : cluster [DBG] pgmap v85: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:57.073 INFO:teuthology.orchestra.run.vm02.stdout:Created osd(s) 1 on host 'vm02' 2026-03-21T06:50:57.157 DEBUG:teuthology.orchestra.run.vm02:osd.1> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.1.service 2026-03-21T06:50:57.158 INFO:tasks.cephadm:Deploying osd.2 on vm04 with /dev/vg_nvme/lv_4... 2026-03-21T06:50:57.158 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- lvm zap /dev/vg_nvme/lv_4 2026-03-21T06:50:57.396 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:50:57 vm02 bash[33726]: debug 2026-03-21T06:50:57.286+0000 7fc6789bd8c0 -1 Falling back to public interface 2026-03-21T06:50:57.399 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.b/config 2026-03-21T06:50:57.896 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:50:57 vm02 bash[33726]: debug 2026-03-21T06:50:57.566+0000 7fc6789bd8c0 -1 osd.1 0 log_to_monitors true 2026-03-21T06:50:57.911 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:50:57.927 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch daemon add osd vm04:vg_nvme/lv_4 --skip-validation 2026-03-21T06:50:58.179 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.b/config 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.063273+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.063273+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.067442+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.067442+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: cluster 2026-03-21T06:50:57.557249+0000 mgr.x (mgr.14152) 140 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: cluster 2026-03-21T06:50:57.557249+0000 mgr.x (mgr.14152) 140 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.574928+0000 mon.a (mon.0) 339 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.574928+0000 mon.a (mon.0) 339 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.582820+0000 mon.b (mon.2) 2 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.582820+0000 mon.b (mon.2) 2 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.820686+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.820686+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.839151+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.839151+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.844309+0000 mon.a (mon.0) 342 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.844309+0000 mon.a (mon.0) 342 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.849719+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.849719+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.850496+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.850496+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.853789+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.853789+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.855936+0000 mon.a (mon.0) 346 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.855936+0000 mon.a (mon.0) 346 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.856347+0000 mon.a (mon.0) 347 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.856347+0000 mon.a (mon.0) 347 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.859652+0000 mon.a (mon.0) 348 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.859652+0000 mon.a (mon.0) 348 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.860609+0000 mon.a (mon.0) 349 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.860609+0000 mon.a (mon.0) 349 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.861676+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.861676+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.862024+0000 mon.a (mon.0) 351 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:58.226 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:58 vm04 bash[20194]: audit 2026-03-21T06:50:57.862024+0000 mon.a (mon.0) 351 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.063273+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.063273+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.067442+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.067442+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: cluster 2026-03-21T06:50:57.557249+0000 mgr.x (mgr.14152) 140 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: cluster 2026-03-21T06:50:57.557249+0000 mgr.x (mgr.14152) 140 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.574928+0000 mon.a (mon.0) 339 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.574928+0000 mon.a (mon.0) 339 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.582820+0000 mon.b (mon.2) 2 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.582820+0000 mon.b (mon.2) 2 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.820686+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.820686+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.839151+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.839151+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.844309+0000 mon.a (mon.0) 342 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.844309+0000 mon.a (mon.0) 342 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.849719+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.849719+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.850496+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.850496+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.853789+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.853789+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.855936+0000 mon.a (mon.0) 346 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.855936+0000 mon.a (mon.0) 346 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.856347+0000 mon.a (mon.0) 347 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.856347+0000 mon.a (mon.0) 347 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.859652+0000 mon.a (mon.0) 348 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.859652+0000 mon.a (mon.0) 348 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.860609+0000 mon.a (mon.0) 349 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.860609+0000 mon.a (mon.0) 349 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.861676+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.861676+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.862024+0000 mon.a (mon.0) 351 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:58.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:58 vm02 bash[17657]: audit 2026-03-21T06:50:57.862024+0000 mon.a (mon.0) 351 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:58.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.063273+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.063273+0000 mon.a (mon.0) 337 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.067442+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.067442+0000 mon.a (mon.0) 338 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: cluster 2026-03-21T06:50:57.557249+0000 mgr.x (mgr.14152) 140 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: cluster 2026-03-21T06:50:57.557249+0000 mgr.x (mgr.14152) 140 : cluster [DBG] pgmap v86: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.574928+0000 mon.a (mon.0) 339 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.574928+0000 mon.a (mon.0) 339 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.582820+0000 mon.b (mon.2) 2 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.582820+0000 mon.b (mon.2) 2 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.820686+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.820686+0000 mon.a (mon.0) 340 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.839151+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.839151+0000 mon.a (mon.0) 341 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.844309+0000 mon.a (mon.0) 342 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.844309+0000 mon.a (mon.0) 342 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.849719+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.849719+0000 mon.a (mon.0) 343 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.850496+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.850496+0000 mon.a (mon.0) 344 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.853789+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.853789+0000 mon.a (mon.0) 345 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.855936+0000 mon.a (mon.0) 346 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.855936+0000 mon.a (mon.0) 346 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.856347+0000 mon.a (mon.0) 347 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.856347+0000 mon.a (mon.0) 347 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.859652+0000 mon.a (mon.0) 348 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.859652+0000 mon.a (mon.0) 348 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.860609+0000 mon.a (mon.0) 349 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.860609+0000 mon.a (mon.0) 349 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.861676+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.861676+0000 mon.a (mon.0) 350 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.862024+0000 mon.a (mon.0) 351 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:58.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:58 vm07 bash[19945]: audit 2026-03-21T06:50:57.862024+0000 mon.a (mon.0) 351 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:59.396 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:50:59 vm02 bash[33726]: debug 2026-03-21T06:50:59.102+0000 7fc67493d640 -1 osd.1 0 waiting for initial osdmap 2026-03-21T06:50:59.396 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:50:59 vm02 bash[33726]: debug 2026-03-21T06:50:59.106+0000 7fc66f724640 -1 osd.1 14 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:57.813979+0000 mgr.x (mgr.14152) 141 : cephadm [INF] Detected new or changed devices on vm02 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:57.813979+0000 mgr.x (mgr.14152) 141 : cephadm [INF] Detected new or changed devices on vm02 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:57.850907+0000 mgr.x (mgr.14152) 142 : cephadm [INF] Adjusting osd_memory_target on vm02 to 1838M 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:57.850907+0000 mgr.x (mgr.14152) 142 : cephadm [INF] Adjusting osd_memory_target on vm02 to 1838M 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.073474+0000 mon.a (mon.0) 352 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.073474+0000 mon.a (mon.0) 352 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cluster 2026-03-21T06:50:58.075408+0000 mon.a (mon.0) 353 : cluster [DBG] osdmap e13: 2 total, 1 up, 2 in 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cluster 2026-03-21T06:50:58.075408+0000 mon.a (mon.0) 353 : cluster [DBG] osdmap e13: 2 total, 1 up, 2 in 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.075569+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.075569+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.076544+0000 mon.a (mon.0) 355 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.076544+0000 mon.a (mon.0) 355 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.084845+0000 mon.b (mon.2) 3 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.084845+0000 mon.b (mon.2) 3 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.533650+0000 mgr.x (mgr.14152) 143 : audit [DBG] from='client.24164 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.533650+0000 mgr.x (mgr.14152) 143 : audit [DBG] from='client.24164 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:58.534443+0000 mgr.x (mgr.14152) 144 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:58.534443+0000 mgr.x (mgr.14152) 144 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:58.534510+0000 mgr.x (mgr.14152) 145 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_4'] 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:58.534510+0000 mgr.x (mgr.14152) 145 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_4'] 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.539650+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.539650+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:58.540090+0000 mgr.x (mgr.14152) 146 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:58.540090+0000 mgr.x (mgr.14152) 146 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:58.540182+0000 mgr.x (mgr.14152) 147 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: cephadm 2026-03-21T06:50:58.540182+0000 mgr.x (mgr.14152) 147 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.543356+0000 mon.a (mon.0) 357 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.543356+0000 mon.a (mon.0) 357 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.544287+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.544287+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.545786+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.545786+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.546220+0000 mon.a (mon.0) 360 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:59.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 bash[17657]: audit 2026-03-21T06:50:58.546220+0000 mon.a (mon.0) 360 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:59.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:57.813979+0000 mgr.x (mgr.14152) 141 : cephadm [INF] Detected new or changed devices on vm02 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:57.813979+0000 mgr.x (mgr.14152) 141 : cephadm [INF] Detected new or changed devices on vm02 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:57.850907+0000 mgr.x (mgr.14152) 142 : cephadm [INF] Adjusting osd_memory_target on vm02 to 1838M 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:57.850907+0000 mgr.x (mgr.14152) 142 : cephadm [INF] Adjusting osd_memory_target on vm02 to 1838M 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.073474+0000 mon.a (mon.0) 352 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.073474+0000 mon.a (mon.0) 352 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cluster 2026-03-21T06:50:58.075408+0000 mon.a (mon.0) 353 : cluster [DBG] osdmap e13: 2 total, 1 up, 2 in 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cluster 2026-03-21T06:50:58.075408+0000 mon.a (mon.0) 353 : cluster [DBG] osdmap e13: 2 total, 1 up, 2 in 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.075569+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.075569+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.076544+0000 mon.a (mon.0) 355 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.076544+0000 mon.a (mon.0) 355 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.084845+0000 mon.b (mon.2) 3 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.084845+0000 mon.b (mon.2) 3 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.533650+0000 mgr.x (mgr.14152) 143 : audit [DBG] from='client.24164 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.533650+0000 mgr.x (mgr.14152) 143 : audit [DBG] from='client.24164 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:58.534443+0000 mgr.x (mgr.14152) 144 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:58.534443+0000 mgr.x (mgr.14152) 144 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:58.534510+0000 mgr.x (mgr.14152) 145 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_4'] 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:58.534510+0000 mgr.x (mgr.14152) 145 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_4'] 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.539650+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.539650+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:58.540090+0000 mgr.x (mgr.14152) 146 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:58.540090+0000 mgr.x (mgr.14152) 146 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:58.540182+0000 mgr.x (mgr.14152) 147 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: cephadm 2026-03-21T06:50:58.540182+0000 mgr.x (mgr.14152) 147 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.543356+0000 mon.a (mon.0) 357 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.543356+0000 mon.a (mon.0) 357 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.544287+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.544287+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.545786+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.545786+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.546220+0000 mon.a (mon.0) 360 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:59.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:50:59 vm07 bash[19945]: audit 2026-03-21T06:50:58.546220+0000 mon.a (mon.0) 360 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:57.813979+0000 mgr.x (mgr.14152) 141 : cephadm [INF] Detected new or changed devices on vm02 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:57.813979+0000 mgr.x (mgr.14152) 141 : cephadm [INF] Detected new or changed devices on vm02 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:57.850907+0000 mgr.x (mgr.14152) 142 : cephadm [INF] Adjusting osd_memory_target on vm02 to 1838M 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:57.850907+0000 mgr.x (mgr.14152) 142 : cephadm [INF] Adjusting osd_memory_target on vm02 to 1838M 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.073474+0000 mon.a (mon.0) 352 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.073474+0000 mon.a (mon.0) 352 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cluster 2026-03-21T06:50:58.075408+0000 mon.a (mon.0) 353 : cluster [DBG] osdmap e13: 2 total, 1 up, 2 in 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cluster 2026-03-21T06:50:58.075408+0000 mon.a (mon.0) 353 : cluster [DBG] osdmap e13: 2 total, 1 up, 2 in 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.075569+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.075569+0000 mon.a (mon.0) 354 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.076544+0000 mon.a (mon.0) 355 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.076544+0000 mon.a (mon.0) 355 : audit [INF] from='osd.1 ' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.084845+0000 mon.b (mon.2) 3 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.084845+0000 mon.b (mon.2) 3 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.533650+0000 mgr.x (mgr.14152) 143 : audit [DBG] from='client.24164 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.533650+0000 mgr.x (mgr.14152) 143 : audit [DBG] from='client.24164 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:58.534443+0000 mgr.x (mgr.14152) 144 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:58.534443+0000 mgr.x (mgr.14152) 144 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:58.534510+0000 mgr.x (mgr.14152) 145 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_4'] 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:58.534510+0000 mgr.x (mgr.14152) 145 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_4'] 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.539650+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.539650+0000 mon.a (mon.0) 356 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:58.540090+0000 mgr.x (mgr.14152) 146 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:58.540090+0000 mgr.x (mgr.14152) 146 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:58.540182+0000 mgr.x (mgr.14152) 147 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: cephadm 2026-03-21T06:50:58.540182+0000 mgr.x (mgr.14152) 147 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.543356+0000 mon.a (mon.0) 357 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.543356+0000 mon.a (mon.0) 357 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.544287+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.544287+0000 mon.a (mon.0) 358 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.545786+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.545786+0000 mon.a (mon.0) 359 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.546220+0000 mon.a (mon.0) 360 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:50:59.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:50:59 vm04 bash[20194]: audit 2026-03-21T06:50:58.546220+0000 mon.a (mon.0) 360 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:00.020 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:50:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:00.020 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:50:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:00.020 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:50:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:00.020 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:50:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:00.020 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:00 vm02 systemd[1]: Stopping Ceph osd.1 for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T06:51:00.396 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:00 vm02 bash[33726]: debug 2026-03-21T06:51:00.054+0000 7fc675952640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-21T06:51:00.396 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:00 vm02 bash[33726]: debug 2026-03-21T06:51:00.054+0000 7fc675952640 -1 osd.1 15 *** Got signal Terminated *** 2026-03-21T06:51:00.396 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:00 vm02 bash[33726]: debug 2026-03-21T06:51:00.054+0000 7fc675952640 -1 osd.1 15 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.076240+0000 mon.a (mon.0) 361 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.076240+0000 mon.a (mon.0) 361 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cluster 2026-03-21T06:50:59.078429+0000 mon.a (mon.0) 362 : cluster [DBG] osdmap e14: 2 total, 1 up, 2 in 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cluster 2026-03-21T06:50:59.078429+0000 mon.a (mon.0) 362 : cluster [DBG] osdmap e14: 2 total, 1 up, 2 in 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.079182+0000 mon.a (mon.0) 363 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.079182+0000 mon.a (mon.0) 363 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.085769+0000 mon.a (mon.0) 364 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.085769+0000 mon.a (mon.0) 364 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.170311+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.170311+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.170897+0000 mon.a (mon.0) 366 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.170897+0000 mon.a (mon.0) 366 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cephadm 2026-03-21T06:50:59.171404+0000 mgr.x (mgr.14152) 148 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cephadm 2026-03-21T06:50:59.171404+0000 mgr.x (mgr.14152) 148 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.447112+0000 mon.a (mon.0) 367 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.447112+0000 mon.a (mon.0) 367 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.450431+0000 mon.a (mon.0) 368 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"}]': finished 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.450431+0000 mon.a (mon.0) 368 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"}]': finished 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cluster 2026-03-21T06:50:59.453778+0000 mon.a (mon.0) 369 : cluster [INF] osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798] boot 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cluster 2026-03-21T06:50:59.453778+0000 mon.a (mon.0) 369 : cluster [INF] osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798] boot 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cluster 2026-03-21T06:50:59.453801+0000 mon.a (mon.0) 370 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cluster 2026-03-21T06:50:59.453801+0000 mon.a (mon.0) 370 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.454445+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.454445+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.454621+0000 mon.a (mon.0) 372 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.454621+0000 mon.a (mon.0) 372 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:00.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.455133+0000 mon.b (mon.2) 4 : audit [INF] from='client.? 192.168.123.104:0/3487529128' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.455133+0000 mon.b (mon.2) 4 : audit [INF] from='client.? 192.168.123.104:0/3487529128' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cluster 2026-03-21T06:50:59.557494+0000 mgr.x (mgr.14152) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:51:00.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cluster 2026-03-21T06:50:59.557494+0000 mgr.x (mgr.14152) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:51:00.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.846612+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.104:0/740738238' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:00.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: audit 2026-03-21T06:50:59.846612+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.104:0/740738238' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:00.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cluster 2026-03-21T06:51:00.056487+0000 mon.a (mon.0) 373 : cluster [INF] osd.1 marked itself down and dead 2026-03-21T06:51:00.397 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 bash[17657]: cluster 2026-03-21T06:51:00.056487+0000 mon.a (mon.0) 373 : cluster [INF] osd.1 marked itself down and dead 2026-03-21T06:51:00.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.076240+0000 mon.a (mon.0) 361 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:51:00.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.076240+0000 mon.a (mon.0) 361 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cluster 2026-03-21T06:50:59.078429+0000 mon.a (mon.0) 362 : cluster [DBG] osdmap e14: 2 total, 1 up, 2 in 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cluster 2026-03-21T06:50:59.078429+0000 mon.a (mon.0) 362 : cluster [DBG] osdmap e14: 2 total, 1 up, 2 in 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.079182+0000 mon.a (mon.0) 363 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.079182+0000 mon.a (mon.0) 363 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.085769+0000 mon.a (mon.0) 364 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.085769+0000 mon.a (mon.0) 364 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.170311+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.170311+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.170897+0000 mon.a (mon.0) 366 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.170897+0000 mon.a (mon.0) 366 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cephadm 2026-03-21T06:50:59.171404+0000 mgr.x (mgr.14152) 148 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cephadm 2026-03-21T06:50:59.171404+0000 mgr.x (mgr.14152) 148 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.447112+0000 mon.a (mon.0) 367 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.447112+0000 mon.a (mon.0) 367 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.450431+0000 mon.a (mon.0) 368 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"}]': finished 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.450431+0000 mon.a (mon.0) 368 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"}]': finished 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cluster 2026-03-21T06:50:59.453778+0000 mon.a (mon.0) 369 : cluster [INF] osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798] boot 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cluster 2026-03-21T06:50:59.453778+0000 mon.a (mon.0) 369 : cluster [INF] osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798] boot 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cluster 2026-03-21T06:50:59.453801+0000 mon.a (mon.0) 370 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cluster 2026-03-21T06:50:59.453801+0000 mon.a (mon.0) 370 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.454445+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.454445+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.454621+0000 mon.a (mon.0) 372 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.454621+0000 mon.a (mon.0) 372 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.455133+0000 mon.b (mon.2) 4 : audit [INF] from='client.? 192.168.123.104:0/3487529128' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.455133+0000 mon.b (mon.2) 4 : audit [INF] from='client.? 192.168.123.104:0/3487529128' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cluster 2026-03-21T06:50:59.557494+0000 mgr.x (mgr.14152) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cluster 2026-03-21T06:50:59.557494+0000 mgr.x (mgr.14152) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.846612+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.104:0/740738238' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: audit 2026-03-21T06:50:59.846612+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.104:0/740738238' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cluster 2026-03-21T06:51:00.056487+0000 mon.a (mon.0) 373 : cluster [INF] osd.1 marked itself down and dead 2026-03-21T06:51:00.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:00 vm07 bash[19945]: cluster 2026-03-21T06:51:00.056487+0000 mon.a (mon.0) 373 : cluster [INF] osd.1 marked itself down and dead 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.076240+0000 mon.a (mon.0) 361 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.076240+0000 mon.a (mon.0) 361 : audit [INF] from='osd.1 ' entity='osd.1' cmd='[{"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]}]': finished 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cluster 2026-03-21T06:50:59.078429+0000 mon.a (mon.0) 362 : cluster [DBG] osdmap e14: 2 total, 1 up, 2 in 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cluster 2026-03-21T06:50:59.078429+0000 mon.a (mon.0) 362 : cluster [DBG] osdmap e14: 2 total, 1 up, 2 in 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.079182+0000 mon.a (mon.0) 363 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.079182+0000 mon.a (mon.0) 363 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.085769+0000 mon.a (mon.0) 364 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.085769+0000 mon.a (mon.0) 364 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.170311+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.170311+0000 mon.a (mon.0) 365 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.1"} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.170897+0000 mon.a (mon.0) 366 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.170897+0000 mon.a (mon.0) 366 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cephadm 2026-03-21T06:50:59.171404+0000 mgr.x (mgr.14152) 148 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cephadm 2026-03-21T06:50:59.171404+0000 mgr.x (mgr.14152) 148 : cephadm [INF] Deploying daemon osd.1 on vm02 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.447112+0000 mon.a (mon.0) 367 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.447112+0000 mon.a (mon.0) 367 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.450431+0000 mon.a (mon.0) 368 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"}]': finished 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.450431+0000 mon.a (mon.0) 368 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"}]': finished 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cluster 2026-03-21T06:50:59.453778+0000 mon.a (mon.0) 369 : cluster [INF] osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798] boot 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cluster 2026-03-21T06:50:59.453778+0000 mon.a (mon.0) 369 : cluster [INF] osd.1 [v2:192.168.123.102:6810/3512774798,v1:192.168.123.102:6811/3512774798] boot 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cluster 2026-03-21T06:50:59.453801+0000 mon.a (mon.0) 370 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cluster 2026-03-21T06:50:59.453801+0000 mon.a (mon.0) 370 : cluster [DBG] osdmap e15: 3 total, 2 up, 3 in 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.454445+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.454445+0000 mon.a (mon.0) 371 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.454621+0000 mon.a (mon.0) 372 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.454621+0000 mon.a (mon.0) 372 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.455133+0000 mon.b (mon.2) 4 : audit [INF] from='client.? 192.168.123.104:0/3487529128' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.455133+0000 mon.b (mon.2) 4 : audit [INF] from='client.? 192.168.123.104:0/3487529128' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "2bbd253a-cfb7-4d35-a80c-5ab1d023b91e"} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cluster 2026-03-21T06:50:59.557494+0000 mgr.x (mgr.14152) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cluster 2026-03-21T06:50:59.557494+0000 mgr.x (mgr.14152) 149 : cluster [DBG] pgmap v90: 0 pgs: ; 0 B data, 27 MiB used, 20 GiB / 20 GiB avail 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.846612+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.104:0/740738238' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: audit 2026-03-21T06:50:59.846612+0000 mon.b (mon.2) 5 : audit [DBG] from='client.? 192.168.123.104:0/740738238' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cluster 2026-03-21T06:51:00.056487+0000 mon.a (mon.0) 373 : cluster [INF] osd.1 marked itself down and dead 2026-03-21T06:51:00.504 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:00 vm04 bash[20194]: cluster 2026-03-21T06:51:00.056487+0000 mon.a (mon.0) 373 : cluster [INF] osd.1 marked itself down and dead 2026-03-21T06:51:00.793 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:00 vm02 bash[37400]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-1 2026-03-21T06:51:01.074 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:51:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:01.075 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:51:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:01.075 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:00 vm02 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.1.service: Deactivated successfully. 2026-03-21T06:51:01.075 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:00 vm02 systemd[1]: Stopped Ceph osd.1 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T06:51:01.075 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:01.075 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:01 vm02 systemd[1]: Started Ceph osd.1 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T06:51:01.075 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:01.338 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:01 vm02 bash[17657]: cluster 2026-03-21T06:51:00.451390+0000 mon.a (mon.0) 374 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:51:01.338 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:01 vm02 bash[17657]: cluster 2026-03-21T06:51:00.451390+0000 mon.a (mon.0) 374 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:51:01.338 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:01 vm02 bash[17657]: cluster 2026-03-21T06:51:00.459488+0000 mon.a (mon.0) 375 : cluster [DBG] osdmap e16: 3 total, 1 up, 3 in 2026-03-21T06:51:01.338 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:01 vm02 bash[17657]: cluster 2026-03-21T06:51:00.459488+0000 mon.a (mon.0) 375 : cluster [DBG] osdmap e16: 3 total, 1 up, 3 in 2026-03-21T06:51:01.338 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:01 vm02 bash[17657]: audit 2026-03-21T06:51:00.459601+0000 mon.a (mon.0) 376 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:01.338 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:01 vm02 bash[17657]: audit 2026-03-21T06:51:00.459601+0000 mon.a (mon.0) 376 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:01.338 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:01 vm02 bash[17657]: audit 2026-03-21T06:51:01.031739+0000 mon.a (mon.0) 377 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:01.338 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:01 vm02 bash[17657]: audit 2026-03-21T06:51:01.031739+0000 mon.a (mon.0) 377 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:01.338 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:01 vm02 bash[17657]: audit 2026-03-21T06:51:01.041551+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:01.338 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:01 vm02 bash[17657]: audit 2026-03-21T06:51:01.041551+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:01.338 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:01 vm02 bash[37622]: Running command: /usr/bin/ceph-authtool --gen-print-key 2026-03-21T06:51:01.338 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:01 vm02 bash[37622]: Running command: /usr/bin/ceph-authtool --gen-print-key 2026-03-21T06:51:01.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:01 vm07 bash[19945]: cluster 2026-03-21T06:51:00.451390+0000 mon.a (mon.0) 374 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:51:01.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:01 vm07 bash[19945]: cluster 2026-03-21T06:51:00.451390+0000 mon.a (mon.0) 374 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:51:01.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:01 vm07 bash[19945]: cluster 2026-03-21T06:51:00.459488+0000 mon.a (mon.0) 375 : cluster [DBG] osdmap e16: 3 total, 1 up, 3 in 2026-03-21T06:51:01.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:01 vm07 bash[19945]: cluster 2026-03-21T06:51:00.459488+0000 mon.a (mon.0) 375 : cluster [DBG] osdmap e16: 3 total, 1 up, 3 in 2026-03-21T06:51:01.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:01 vm07 bash[19945]: audit 2026-03-21T06:51:00.459601+0000 mon.a (mon.0) 376 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:01.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:01 vm07 bash[19945]: audit 2026-03-21T06:51:00.459601+0000 mon.a (mon.0) 376 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:01.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:01 vm07 bash[19945]: audit 2026-03-21T06:51:01.031739+0000 mon.a (mon.0) 377 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:01.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:01 vm07 bash[19945]: audit 2026-03-21T06:51:01.031739+0000 mon.a (mon.0) 377 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:01.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:01 vm07 bash[19945]: audit 2026-03-21T06:51:01.041551+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:01.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:01 vm07 bash[19945]: audit 2026-03-21T06:51:01.041551+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:01.429 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:01 vm04 bash[20194]: cluster 2026-03-21T06:51:00.451390+0000 mon.a (mon.0) 374 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:51:01.429 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:01 vm04 bash[20194]: cluster 2026-03-21T06:51:00.451390+0000 mon.a (mon.0) 374 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:51:01.429 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:01 vm04 bash[20194]: cluster 2026-03-21T06:51:00.459488+0000 mon.a (mon.0) 375 : cluster [DBG] osdmap e16: 3 total, 1 up, 3 in 2026-03-21T06:51:01.429 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:01 vm04 bash[20194]: cluster 2026-03-21T06:51:00.459488+0000 mon.a (mon.0) 375 : cluster [DBG] osdmap e16: 3 total, 1 up, 3 in 2026-03-21T06:51:01.429 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:01 vm04 bash[20194]: audit 2026-03-21T06:51:00.459601+0000 mon.a (mon.0) 376 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:01.429 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:01 vm04 bash[20194]: audit 2026-03-21T06:51:00.459601+0000 mon.a (mon.0) 376 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:01.429 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:01 vm04 bash[20194]: audit 2026-03-21T06:51:01.031739+0000 mon.a (mon.0) 377 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:01.429 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:01 vm04 bash[20194]: audit 2026-03-21T06:51:01.031739+0000 mon.a (mon.0) 377 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:01.429 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:01 vm04 bash[20194]: audit 2026-03-21T06:51:01.041551+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:01.429 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:01 vm04 bash[20194]: audit 2026-03-21T06:51:01.041551+0000 mon.a (mon.0) 378 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:02.279 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:02 vm04 bash[20194]: audit 2026-03-21T06:51:01.467607+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch 2026-03-21T06:51:02.280 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:02 vm04 bash[20194]: audit 2026-03-21T06:51:01.467607+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch 2026-03-21T06:51:02.280 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:02 vm04 bash[20194]: audit 2026-03-21T06:51:01.471244+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:02.280 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:02 vm04 bash[20194]: audit 2026-03-21T06:51:01.471244+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:02.280 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:02 vm04 bash[20194]: cephadm 2026-03-21T06:51:01.473329+0000 mgr.x (mgr.14152) 150 : cephadm [INF] Deploying daemon osd.2 on vm04 2026-03-21T06:51:02.280 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:02 vm04 bash[20194]: cephadm 2026-03-21T06:51:01.473329+0000 mgr.x (mgr.14152) 150 : cephadm [INF] Deploying daemon osd.2 on vm04 2026-03-21T06:51:02.280 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:02 vm04 bash[20194]: cluster 2026-03-21T06:51:01.557748+0000 mgr.x (mgr.14152) 151 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:02.280 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:02 vm04 bash[20194]: cluster 2026-03-21T06:51:01.557748+0000 mgr.x (mgr.14152) 151 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:02.280 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:02 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:02.362 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:02 vm02 bash[17657]: audit 2026-03-21T06:51:01.467607+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch 2026-03-21T06:51:02.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:02 vm02 bash[17657]: audit 2026-03-21T06:51:01.467607+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch 2026-03-21T06:51:02.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:02 vm02 bash[17657]: audit 2026-03-21T06:51:01.471244+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:02.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:02 vm02 bash[17657]: audit 2026-03-21T06:51:01.471244+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:02.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:02 vm02 bash[17657]: cephadm 2026-03-21T06:51:01.473329+0000 mgr.x (mgr.14152) 150 : cephadm [INF] Deploying daemon osd.2 on vm04 2026-03-21T06:51:02.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:02 vm02 bash[17657]: cephadm 2026-03-21T06:51:01.473329+0000 mgr.x (mgr.14152) 150 : cephadm [INF] Deploying daemon osd.2 on vm04 2026-03-21T06:51:02.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:02 vm02 bash[17657]: cluster 2026-03-21T06:51:01.557748+0000 mgr.x (mgr.14152) 151 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:02.363 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:02 vm02 bash[17657]: cluster 2026-03-21T06:51:01.557748+0000 mgr.x (mgr.14152) 151 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:02.363 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:02 vm02 bash[37622]: --> Failed to activate via raw: did not find any matching OSD to activate 2026-03-21T06:51:02.363 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:02 vm02 bash[37622]: Running command: /usr/bin/ceph-authtool --gen-print-key 2026-03-21T06:51:02.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:02 vm07 bash[19945]: audit 2026-03-21T06:51:01.467607+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch 2026-03-21T06:51:02.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:02 vm07 bash[19945]: audit 2026-03-21T06:51:01.467607+0000 mon.a (mon.0) 379 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.2"} : dispatch 2026-03-21T06:51:02.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:02 vm07 bash[19945]: audit 2026-03-21T06:51:01.471244+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:02.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:02 vm07 bash[19945]: audit 2026-03-21T06:51:01.471244+0000 mon.a (mon.0) 380 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:02.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:02 vm07 bash[19945]: cephadm 2026-03-21T06:51:01.473329+0000 mgr.x (mgr.14152) 150 : cephadm [INF] Deploying daemon osd.2 on vm04 2026-03-21T06:51:02.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:02 vm07 bash[19945]: cephadm 2026-03-21T06:51:01.473329+0000 mgr.x (mgr.14152) 150 : cephadm [INF] Deploying daemon osd.2 on vm04 2026-03-21T06:51:02.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:02 vm07 bash[19945]: cluster 2026-03-21T06:51:01.557748+0000 mgr.x (mgr.14152) 151 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:02.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:02 vm07 bash[19945]: cluster 2026-03-21T06:51:01.557748+0000 mgr.x (mgr.14152) 151 : cluster [DBG] pgmap v92: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:02.549 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:02 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:02.646 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:02 vm02 bash[37622]: Running command: /usr/bin/ceph-authtool --gen-print-key 2026-03-21T06:51:02.646 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:02 vm02 bash[37622]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1 2026-03-21T06:51:02.646 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:02 vm02 bash[37622]: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/vg_nvme/lv_3 --path /var/lib/ceph/osd/ceph-1 --no-mon-config 2026-03-21T06:51:02.646 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:02 vm02 bash[37622]: Running command: /usr/bin/ln -snf /dev/vg_nvme/lv_3 /var/lib/ceph/osd/ceph-1/block 2026-03-21T06:51:02.646 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:02 vm02 bash[37622]: Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-1/block 2026-03-21T06:51:02.646 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:02 vm02 bash[37622]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-2 2026-03-21T06:51:02.646 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:02 vm02 bash[37622]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-1 2026-03-21T06:51:02.646 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:02 vm02 bash[37622]: --> ceph-volume lvm activate successful for osd ID: 1 2026-03-21T06:51:03.050 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:02 vm02 bash[38374]: debug 2026-03-21T06:51:02.754+0000 7f702c424640 1 -- 192.168.123.102:0/86390821 <== mon.0 v2:192.168.123.102:3300/0 4 ==== auth_reply(proto 2 0 (0) Success) ==== 194+0+0 (secure 0 0 0) 0x55dd5d667860 con 0x55dd5959dc00 2026-03-21T06:51:03.646 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:03 vm02 bash[38374]: debug 2026-03-21T06:51:03.042+0000 7f702ec8e8c0 -1 Falling back to public interface 2026-03-21T06:51:03.646 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:03 vm02 bash[38374]: debug 2026-03-21T06:51:03.314+0000 7f702ec8e8c0 -1 osd.1 15 log_to_monitors true 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:02.494277+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:02.494277+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:02.500264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:02.500264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:02.612833+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:02.612833+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:02.617147+0000 mon.a (mon.0) 384 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:02.617147+0000 mon.a (mon.0) 384 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:02.632333+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:02.632333+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:03.319428+0000 mon.a (mon.0) 386 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:51:03.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:03 vm02 bash[17657]: audit 2026-03-21T06:51:03.319428+0000 mon.a (mon.0) 386 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:51:03.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:02.494277+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:02.494277+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:02.500264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:02.500264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:02.612833+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:02.612833+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:02.617147+0000 mon.a (mon.0) 384 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:02.617147+0000 mon.a (mon.0) 384 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.753 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:02.632333+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:03.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:02.632333+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:03.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:03.319428+0000 mon.a (mon.0) 386 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:51:03.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:03 vm04 bash[20194]: audit 2026-03-21T06:51:03.319428+0000 mon.a (mon.0) 386 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:51:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:02.494277+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:02.494277+0000 mon.a (mon.0) 381 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:02.500264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:02.500264+0000 mon.a (mon.0) 382 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:02.612833+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:02.612833+0000 mon.a (mon.0) 383 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:02.617147+0000 mon.a (mon.0) 384 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:02.617147+0000 mon.a (mon.0) 384 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:02.632333+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:02.632333+0000 mon.a (mon.0) 385 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:03.319428+0000 mon.a (mon.0) 386 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:51:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:03 vm07 bash[19945]: audit 2026-03-21T06:51:03.319428+0000 mon.a (mon.0) 386 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]} : dispatch 2026-03-21T06:51:04.146 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:51:03 vm02 bash[38374]: debug 2026-03-21T06:51:03.646+0000 7f70259f5640 -1 osd.1 15 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-21T06:51:04.355 INFO:teuthology.orchestra.run.vm04.stdout:Created osd(s) 2 on host 'vm04' 2026-03-21T06:51:04.426 DEBUG:teuthology.orchestra.run.vm04:osd.2> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.2.service 2026-03-21T06:51:04.427 INFO:tasks.cephadm:Deploying osd.3 on vm04 with /dev/vg_nvme/lv_3... 2026-03-21T06:51:04.427 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- lvm zap /dev/vg_nvme/lv_3 2026-03-21T06:51:04.695 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: cluster 2026-03-21T06:51:03.558071+0000 mgr.x (mgr.14152) 152 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:04.695 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: cluster 2026-03-21T06:51:03.558071+0000 mgr.x (mgr.14152) 152 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:04.695 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: audit 2026-03-21T06:51:03.619040+0000 mon.a (mon.0) 387 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:51:04.695 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: audit 2026-03-21T06:51:03.619040+0000 mon.a (mon.0) 387 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:51:04.696 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: cluster 2026-03-21T06:51:03.621314+0000 mon.a (mon.0) 388 : cluster [DBG] osdmap e17: 3 total, 1 up, 3 in 2026-03-21T06:51:04.696 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: cluster 2026-03-21T06:51:03.621314+0000 mon.a (mon.0) 388 : cluster [DBG] osdmap e17: 3 total, 1 up, 3 in 2026-03-21T06:51:04.696 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: audit 2026-03-21T06:51:03.621546+0000 mon.a (mon.0) 389 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:51:04.696 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: audit 2026-03-21T06:51:03.621546+0000 mon.a (mon.0) 389 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:51:04.696 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: audit 2026-03-21T06:51:03.621728+0000 mon.a (mon.0) 390 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:04.696 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: audit 2026-03-21T06:51:03.621728+0000 mon.a (mon.0) 390 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:04.696 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: audit 2026-03-21T06:51:04.343410+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:04.696 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: audit 2026-03-21T06:51:04.343410+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:04.696 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: audit 2026-03-21T06:51:04.347517+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:04.696 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:04 vm04 bash[20194]: audit 2026-03-21T06:51:04.347517+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:04.720 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.b/config 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: cluster 2026-03-21T06:51:03.558071+0000 mgr.x (mgr.14152) 152 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: cluster 2026-03-21T06:51:03.558071+0000 mgr.x (mgr.14152) 152 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: audit 2026-03-21T06:51:03.619040+0000 mon.a (mon.0) 387 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: audit 2026-03-21T06:51:03.619040+0000 mon.a (mon.0) 387 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: cluster 2026-03-21T06:51:03.621314+0000 mon.a (mon.0) 388 : cluster [DBG] osdmap e17: 3 total, 1 up, 3 in 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: cluster 2026-03-21T06:51:03.621314+0000 mon.a (mon.0) 388 : cluster [DBG] osdmap e17: 3 total, 1 up, 3 in 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: audit 2026-03-21T06:51:03.621546+0000 mon.a (mon.0) 389 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: audit 2026-03-21T06:51:03.621546+0000 mon.a (mon.0) 389 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: audit 2026-03-21T06:51:03.621728+0000 mon.a (mon.0) 390 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: audit 2026-03-21T06:51:03.621728+0000 mon.a (mon.0) 390 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: audit 2026-03-21T06:51:04.343410+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:04.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: audit 2026-03-21T06:51:04.343410+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:04.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: audit 2026-03-21T06:51:04.347517+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:04.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:04 vm02 bash[17657]: audit 2026-03-21T06:51:04.347517+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: cluster 2026-03-21T06:51:03.558071+0000 mgr.x (mgr.14152) 152 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: cluster 2026-03-21T06:51:03.558071+0000 mgr.x (mgr.14152) 152 : cluster [DBG] pgmap v93: 0 pgs: ; 0 B data, 53 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: audit 2026-03-21T06:51:03.619040+0000 mon.a (mon.0) 387 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:51:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: audit 2026-03-21T06:51:03.619040+0000 mon.a (mon.0) 387 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["1"]}]': finished 2026-03-21T06:51:04.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: cluster 2026-03-21T06:51:03.621314+0000 mon.a (mon.0) 388 : cluster [DBG] osdmap e17: 3 total, 1 up, 3 in 2026-03-21T06:51:04.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: cluster 2026-03-21T06:51:03.621314+0000 mon.a (mon.0) 388 : cluster [DBG] osdmap e17: 3 total, 1 up, 3 in 2026-03-21T06:51:04.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: audit 2026-03-21T06:51:03.621546+0000 mon.a (mon.0) 389 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:51:04.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: audit 2026-03-21T06:51:03.621546+0000 mon.a (mon.0) 389 : audit [INF] from='osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959]' entity='osd.1' cmd={"prefix": "osd crush create-or-move", "id": 1, "weight":0.0195, "args": ["host=vm02", "root=default"]} : dispatch 2026-03-21T06:51:04.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: audit 2026-03-21T06:51:03.621728+0000 mon.a (mon.0) 390 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:04.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: audit 2026-03-21T06:51:03.621728+0000 mon.a (mon.0) 390 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:04.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: audit 2026-03-21T06:51:04.343410+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:04.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: audit 2026-03-21T06:51:04.343410+0000 mon.a (mon.0) 391 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:04.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: audit 2026-03-21T06:51:04.347517+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:04.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:04 vm07 bash[19945]: audit 2026-03-21T06:51:04.347517+0000 mon.a (mon.0) 392 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:05.003 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:51:04 vm04 bash[23182]: debug 2026-03-21T06:51:04.695+0000 7f3b5d1b78c0 -1 osd.2 0 log_to_monitors true 2026-03-21T06:51:05.189 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:51:05.204 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch daemon add osd vm04:vg_nvme/lv_3 --skip-validation 2026-03-21T06:51:05.433 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.b/config 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: cluster 2026-03-21T06:51:04.619231+0000 mon.a (mon.0) 393 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: cluster 2026-03-21T06:51:04.619231+0000 mon.a (mon.0) 393 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: cluster 2026-03-21T06:51:04.619260+0000 mon.a (mon.0) 394 : cluster [INF] Cluster is now healthy 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: cluster 2026-03-21T06:51:04.619260+0000 mon.a (mon.0) 394 : cluster [INF] Cluster is now healthy 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: cluster 2026-03-21T06:51:04.624104+0000 mon.a (mon.0) 395 : cluster [INF] osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959] boot 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: cluster 2026-03-21T06:51:04.624104+0000 mon.a (mon.0) 395 : cluster [INF] osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959] boot 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: cluster 2026-03-21T06:51:04.624148+0000 mon.a (mon.0) 396 : cluster [DBG] osdmap e18: 3 total, 2 up, 3 in 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: cluster 2026-03-21T06:51:04.624148+0000 mon.a (mon.0) 396 : cluster [DBG] osdmap e18: 3 total, 2 up, 3 in 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: audit 2026-03-21T06:51:04.624996+0000 mon.a (mon.0) 397 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: audit 2026-03-21T06:51:04.624996+0000 mon.a (mon.0) 397 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: audit 2026-03-21T06:51:04.625141+0000 mon.a (mon.0) 398 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: audit 2026-03-21T06:51:04.625141+0000 mon.a (mon.0) 398 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: audit 2026-03-21T06:51:04.694259+0000 mon.a (mon.0) 399 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: audit 2026-03-21T06:51:04.694259+0000 mon.a (mon.0) 399 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: audit 2026-03-21T06:51:04.702412+0000 mon.b (mon.2) 6 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:05.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:05 vm04 bash[20194]: audit 2026-03-21T06:51:04.702412+0000 mon.b (mon.2) 6 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: cluster 2026-03-21T06:51:04.619231+0000 mon.a (mon.0) 393 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: cluster 2026-03-21T06:51:04.619231+0000 mon.a (mon.0) 393 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: cluster 2026-03-21T06:51:04.619260+0000 mon.a (mon.0) 394 : cluster [INF] Cluster is now healthy 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: cluster 2026-03-21T06:51:04.619260+0000 mon.a (mon.0) 394 : cluster [INF] Cluster is now healthy 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: cluster 2026-03-21T06:51:04.624104+0000 mon.a (mon.0) 395 : cluster [INF] osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959] boot 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: cluster 2026-03-21T06:51:04.624104+0000 mon.a (mon.0) 395 : cluster [INF] osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959] boot 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: cluster 2026-03-21T06:51:04.624148+0000 mon.a (mon.0) 396 : cluster [DBG] osdmap e18: 3 total, 2 up, 3 in 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: cluster 2026-03-21T06:51:04.624148+0000 mon.a (mon.0) 396 : cluster [DBG] osdmap e18: 3 total, 2 up, 3 in 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: audit 2026-03-21T06:51:04.624996+0000 mon.a (mon.0) 397 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: audit 2026-03-21T06:51:04.624996+0000 mon.a (mon.0) 397 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: audit 2026-03-21T06:51:04.625141+0000 mon.a (mon.0) 398 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: audit 2026-03-21T06:51:04.625141+0000 mon.a (mon.0) 398 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: audit 2026-03-21T06:51:04.694259+0000 mon.a (mon.0) 399 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: audit 2026-03-21T06:51:04.694259+0000 mon.a (mon.0) 399 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: audit 2026-03-21T06:51:04.702412+0000 mon.b (mon.2) 6 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:05.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:05 vm02 bash[17657]: audit 2026-03-21T06:51:04.702412+0000 mon.b (mon.2) 6 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: cluster 2026-03-21T06:51:04.619231+0000 mon.a (mon.0) 393 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:51:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: cluster 2026-03-21T06:51:04.619231+0000 mon.a (mon.0) 393 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:51:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: cluster 2026-03-21T06:51:04.619260+0000 mon.a (mon.0) 394 : cluster [INF] Cluster is now healthy 2026-03-21T06:51:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: cluster 2026-03-21T06:51:04.619260+0000 mon.a (mon.0) 394 : cluster [INF] Cluster is now healthy 2026-03-21T06:51:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: cluster 2026-03-21T06:51:04.624104+0000 mon.a (mon.0) 395 : cluster [INF] osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959] boot 2026-03-21T06:51:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: cluster 2026-03-21T06:51:04.624104+0000 mon.a (mon.0) 395 : cluster [INF] osd.1 [v2:192.168.123.102:6810/2868931959,v1:192.168.123.102:6811/2868931959] boot 2026-03-21T06:51:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: cluster 2026-03-21T06:51:04.624148+0000 mon.a (mon.0) 396 : cluster [DBG] osdmap e18: 3 total, 2 up, 3 in 2026-03-21T06:51:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: cluster 2026-03-21T06:51:04.624148+0000 mon.a (mon.0) 396 : cluster [DBG] osdmap e18: 3 total, 2 up, 3 in 2026-03-21T06:51:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: audit 2026-03-21T06:51:04.624996+0000 mon.a (mon.0) 397 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: audit 2026-03-21T06:51:04.624996+0000 mon.a (mon.0) 397 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 1} : dispatch 2026-03-21T06:51:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: audit 2026-03-21T06:51:04.625141+0000 mon.a (mon.0) 398 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:05.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: audit 2026-03-21T06:51:04.625141+0000 mon.a (mon.0) 398 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:05.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: audit 2026-03-21T06:51:04.694259+0000 mon.a (mon.0) 399 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:05.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: audit 2026-03-21T06:51:04.694259+0000 mon.a (mon.0) 399 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:05.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: audit 2026-03-21T06:51:04.702412+0000 mon.b (mon.2) 6 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:05.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:05 vm07 bash[19945]: audit 2026-03-21T06:51:04.702412+0000 mon.b (mon.2) 6 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]} : dispatch 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: cluster 2026-03-21T06:51:05.558321+0000 mgr.x (mgr.14152) 153 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: cluster 2026-03-21T06:51:05.558321+0000 mgr.x (mgr.14152) 153 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.634725+0000 mon.a (mon.0) 400 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.634725+0000 mon.a (mon.0) 400 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: cluster 2026-03-21T06:51:05.636323+0000 mon.a (mon.0) 401 : cluster [DBG] osdmap e19: 3 total, 2 up, 3 in 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: cluster 2026-03-21T06:51:05.636323+0000 mon.a (mon.0) 401 : cluster [DBG] osdmap e19: 3 total, 2 up, 3 in 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.636528+0000 mon.a (mon.0) 402 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.636528+0000 mon.a (mon.0) 402 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.637626+0000 mon.a (mon.0) 403 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.637626+0000 mon.a (mon.0) 403 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.645917+0000 mon.b (mon.2) 7 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.645917+0000 mon.b (mon.2) 7 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.757809+0000 mon.a (mon.0) 404 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:06.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.757809+0000 mon.a (mon.0) 404 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.761028+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.761028+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.761794+0000 mon.a (mon.0) 406 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.761794+0000 mon.a (mon.0) 406 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.763425+0000 mon.a (mon.0) 407 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.763425+0000 mon.a (mon.0) 407 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.764023+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:05.764023+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.602169+0000 mon.a (mon.0) 409 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.602169+0000 mon.a (mon.0) 409 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.605349+0000 mon.a (mon.0) 410 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.605349+0000 mon.a (mon.0) 410 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.605412+0000 mon.a (mon.0) 411 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"}]': finished 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.605412+0000 mon.a (mon.0) 411 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"}]': finished 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: cluster 2026-03-21T06:51:06.607051+0000 mon.a (mon.0) 412 : cluster [DBG] osdmap e20: 4 total, 2 up, 4 in 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: cluster 2026-03-21T06:51:06.607051+0000 mon.a (mon.0) 412 : cluster [DBG] osdmap e20: 4 total, 2 up, 4 in 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.607169+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.607169+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.607253+0000 mon.a (mon.0) 414 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.607253+0000 mon.a (mon.0) 414 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.610297+0000 mon.b (mon.2) 8 : audit [INF] from='client.? 192.168.123.104:0/2832590660' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.610297+0000 mon.b (mon.2) 8 : audit [INF] from='client.? 192.168.123.104:0/2832590660' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.611623+0000 mon.a (mon.0) 415 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:06.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:06 vm02 bash[17657]: audit 2026-03-21T06:51:06.611623+0000 mon.a (mon.0) 415 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: cluster 2026-03-21T06:51:05.558321+0000 mgr.x (mgr.14152) 153 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: cluster 2026-03-21T06:51:05.558321+0000 mgr.x (mgr.14152) 153 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.634725+0000 mon.a (mon.0) 400 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-21T06:51:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.634725+0000 mon.a (mon.0) 400 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-21T06:51:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: cluster 2026-03-21T06:51:05.636323+0000 mon.a (mon.0) 401 : cluster [DBG] osdmap e19: 3 total, 2 up, 3 in 2026-03-21T06:51:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: cluster 2026-03-21T06:51:05.636323+0000 mon.a (mon.0) 401 : cluster [DBG] osdmap e19: 3 total, 2 up, 3 in 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.636528+0000 mon.a (mon.0) 402 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.636528+0000 mon.a (mon.0) 402 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.637626+0000 mon.a (mon.0) 403 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.637626+0000 mon.a (mon.0) 403 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.645917+0000 mon.b (mon.2) 7 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.645917+0000 mon.b (mon.2) 7 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.757809+0000 mon.a (mon.0) 404 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.757809+0000 mon.a (mon.0) 404 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.761028+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.761028+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.761794+0000 mon.a (mon.0) 406 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.761794+0000 mon.a (mon.0) 406 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.763425+0000 mon.a (mon.0) 407 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.763425+0000 mon.a (mon.0) 407 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.764023+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:05.764023+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.602169+0000 mon.a (mon.0) 409 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.602169+0000 mon.a (mon.0) 409 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.605349+0000 mon.a (mon.0) 410 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.605349+0000 mon.a (mon.0) 410 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.605412+0000 mon.a (mon.0) 411 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"}]': finished 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.605412+0000 mon.a (mon.0) 411 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"}]': finished 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: cluster 2026-03-21T06:51:06.607051+0000 mon.a (mon.0) 412 : cluster [DBG] osdmap e20: 4 total, 2 up, 4 in 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: cluster 2026-03-21T06:51:06.607051+0000 mon.a (mon.0) 412 : cluster [DBG] osdmap e20: 4 total, 2 up, 4 in 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.607169+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.607169+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.607253+0000 mon.a (mon.0) 414 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.607253+0000 mon.a (mon.0) 414 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.610297+0000 mon.b (mon.2) 8 : audit [INF] from='client.? 192.168.123.104:0/2832590660' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.610297+0000 mon.b (mon.2) 8 : audit [INF] from='client.? 192.168.123.104:0/2832590660' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.611623+0000 mon.a (mon.0) 415 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:06 vm07 bash[19945]: audit 2026-03-21T06:51:06.611623+0000 mon.a (mon.0) 415 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.003 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:51:06 vm04 bash[23182]: debug 2026-03-21T06:51:06.631+0000 7f3b59137640 -1 osd.2 0 waiting for initial osdmap 2026-03-21T06:51:07.003 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:51:06 vm04 bash[23182]: debug 2026-03-21T06:51:06.635+0000 7f3b53f1e640 -1 osd.2 20 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: cluster 2026-03-21T06:51:05.558321+0000 mgr.x (mgr.14152) 153 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: cluster 2026-03-21T06:51:05.558321+0000 mgr.x (mgr.14152) 153 : cluster [DBG] pgmap v96: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.634725+0000 mon.a (mon.0) 400 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.634725+0000 mon.a (mon.0) 400 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["2"]}]': finished 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: cluster 2026-03-21T06:51:05.636323+0000 mon.a (mon.0) 401 : cluster [DBG] osdmap e19: 3 total, 2 up, 3 in 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: cluster 2026-03-21T06:51:05.636323+0000 mon.a (mon.0) 401 : cluster [DBG] osdmap e19: 3 total, 2 up, 3 in 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.636528+0000 mon.a (mon.0) 402 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.636528+0000 mon.a (mon.0) 402 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.637626+0000 mon.a (mon.0) 403 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.637626+0000 mon.a (mon.0) 403 : audit [INF] from='osd.2 ' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.645917+0000 mon.b (mon.2) 7 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.645917+0000 mon.b (mon.2) 7 : audit [INF] from='osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046]' entity='osd.2' cmd={"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.757809+0000 mon.a (mon.0) 404 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.757809+0000 mon.a (mon.0) 404 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.761028+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.761028+0000 mon.a (mon.0) 405 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.761794+0000 mon.a (mon.0) 406 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.761794+0000 mon.a (mon.0) 406 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.763425+0000 mon.a (mon.0) 407 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.763425+0000 mon.a (mon.0) 407 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.764023+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:05.764023+0000 mon.a (mon.0) 408 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.602169+0000 mon.a (mon.0) 409 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.602169+0000 mon.a (mon.0) 409 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.605349+0000 mon.a (mon.0) 410 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.605349+0000 mon.a (mon.0) 410 : audit [INF] from='osd.2 ' entity='osd.2' cmd='[{"prefix": "osd crush create-or-move", "id": 2, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.605412+0000 mon.a (mon.0) 411 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"}]': finished 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.605412+0000 mon.a (mon.0) 411 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"}]': finished 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: cluster 2026-03-21T06:51:06.607051+0000 mon.a (mon.0) 412 : cluster [DBG] osdmap e20: 4 total, 2 up, 4 in 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: cluster 2026-03-21T06:51:06.607051+0000 mon.a (mon.0) 412 : cluster [DBG] osdmap e20: 4 total, 2 up, 4 in 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.607169+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.607169+0000 mon.a (mon.0) 413 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.607253+0000 mon.a (mon.0) 414 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.607253+0000 mon.a (mon.0) 414 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.610297+0000 mon.b (mon.2) 8 : audit [INF] from='client.? 192.168.123.104:0/2832590660' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.610297+0000 mon.b (mon.2) 8 : audit [INF] from='client.? 192.168.123.104:0/2832590660' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "c2851fc5-1154-4c44-91cd-287a260b97a7"} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.611623+0000 mon.a (mon.0) 415 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:06 vm04 bash[20194]: audit 2026-03-21T06:51:06.611623+0000 mon.a (mon.0) 415 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:05.752850+0000 mgr.x (mgr.14152) 154 : audit [DBG] from='client.24197 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:05.752850+0000 mgr.x (mgr.14152) 154 : audit [DBG] from='client.24197 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cephadm 2026-03-21T06:51:05.753679+0000 mgr.x (mgr.14152) 155 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cephadm 2026-03-21T06:51:05.753679+0000 mgr.x (mgr.14152) 155 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cephadm 2026-03-21T06:51:05.753773+0000 mgr.x (mgr.14152) 156 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_3'] 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cephadm 2026-03-21T06:51:05.753773+0000 mgr.x (mgr.14152) 156 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_3'] 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cephadm 2026-03-21T06:51:05.758132+0000 mgr.x (mgr.14152) 157 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cephadm 2026-03-21T06:51:05.758132+0000 mgr.x (mgr.14152) 157 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cephadm 2026-03-21T06:51:05.758226+0000 mgr.x (mgr.14152) 158 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cephadm 2026-03-21T06:51:05.758226+0000 mgr.x (mgr.14152) 158 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:06.961673+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.104:0/583254995' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:06.961673+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.104:0/583254995' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cluster 2026-03-21T06:51:07.612535+0000 mon.a (mon.0) 416 : cluster [INF] osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046] boot 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cluster 2026-03-21T06:51:07.612535+0000 mon.a (mon.0) 416 : cluster [INF] osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046] boot 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cluster 2026-03-21T06:51:07.612572+0000 mon.a (mon.0) 417 : cluster [DBG] osdmap e21: 4 total, 3 up, 4 in 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: cluster 2026-03-21T06:51:07.612572+0000 mon.a (mon.0) 417 : cluster [DBG] osdmap e21: 4 total, 3 up, 4 in 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:07.612694+0000 mon.a (mon.0) 418 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:07.612694+0000 mon.a (mon.0) 418 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:07.612861+0000 mon.a (mon.0) 419 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:07.612861+0000 mon.a (mon.0) 419 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:07.623657+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:07.623657+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:07.629019+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:07 vm07 bash[19945]: audit 2026-03-21T06:51:07.629019+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:05.752850+0000 mgr.x (mgr.14152) 154 : audit [DBG] from='client.24197 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:05.752850+0000 mgr.x (mgr.14152) 154 : audit [DBG] from='client.24197 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cephadm 2026-03-21T06:51:05.753679+0000 mgr.x (mgr.14152) 155 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cephadm 2026-03-21T06:51:05.753679+0000 mgr.x (mgr.14152) 155 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cephadm 2026-03-21T06:51:05.753773+0000 mgr.x (mgr.14152) 156 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_3'] 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cephadm 2026-03-21T06:51:05.753773+0000 mgr.x (mgr.14152) 156 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_3'] 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cephadm 2026-03-21T06:51:05.758132+0000 mgr.x (mgr.14152) 157 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cephadm 2026-03-21T06:51:05.758132+0000 mgr.x (mgr.14152) 157 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cephadm 2026-03-21T06:51:05.758226+0000 mgr.x (mgr.14152) 158 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cephadm 2026-03-21T06:51:05.758226+0000 mgr.x (mgr.14152) 158 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:06.961673+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.104:0/583254995' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:06.961673+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.104:0/583254995' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cluster 2026-03-21T06:51:07.612535+0000 mon.a (mon.0) 416 : cluster [INF] osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046] boot 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cluster 2026-03-21T06:51:07.612535+0000 mon.a (mon.0) 416 : cluster [INF] osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046] boot 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cluster 2026-03-21T06:51:07.612572+0000 mon.a (mon.0) 417 : cluster [DBG] osdmap e21: 4 total, 3 up, 4 in 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: cluster 2026-03-21T06:51:07.612572+0000 mon.a (mon.0) 417 : cluster [DBG] osdmap e21: 4 total, 3 up, 4 in 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:07.612694+0000 mon.a (mon.0) 418 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:07.612694+0000 mon.a (mon.0) 418 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:07.612861+0000 mon.a (mon.0) 419 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:07.612861+0000 mon.a (mon.0) 419 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:07.623657+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:07.623657+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:07.629019+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.922 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:07 vm02 bash[17657]: audit 2026-03-21T06:51:07.629019+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:05.752850+0000 mgr.x (mgr.14152) 154 : audit [DBG] from='client.24197 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:05.752850+0000 mgr.x (mgr.14152) 154 : audit [DBG] from='client.24197 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cephadm 2026-03-21T06:51:05.753679+0000 mgr.x (mgr.14152) 155 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cephadm 2026-03-21T06:51:05.753679+0000 mgr.x (mgr.14152) 155 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cephadm 2026-03-21T06:51:05.753773+0000 mgr.x (mgr.14152) 156 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_3'] 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cephadm 2026-03-21T06:51:05.753773+0000 mgr.x (mgr.14152) 156 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_3'] 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cephadm 2026-03-21T06:51:05.758132+0000 mgr.x (mgr.14152) 157 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cephadm 2026-03-21T06:51:05.758132+0000 mgr.x (mgr.14152) 157 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cephadm 2026-03-21T06:51:05.758226+0000 mgr.x (mgr.14152) 158 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cephadm 2026-03-21T06:51:05.758226+0000 mgr.x (mgr.14152) 158 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:06.961673+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.104:0/583254995' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:06.961673+0000 mon.b (mon.2) 9 : audit [DBG] from='client.? 192.168.123.104:0/583254995' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cluster 2026-03-21T06:51:07.612535+0000 mon.a (mon.0) 416 : cluster [INF] osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046] boot 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cluster 2026-03-21T06:51:07.612535+0000 mon.a (mon.0) 416 : cluster [INF] osd.2 [v2:192.168.123.104:6800/3750025046,v1:192.168.123.104:6801/3750025046] boot 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cluster 2026-03-21T06:51:07.612572+0000 mon.a (mon.0) 417 : cluster [DBG] osdmap e21: 4 total, 3 up, 4 in 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: cluster 2026-03-21T06:51:07.612572+0000 mon.a (mon.0) 417 : cluster [DBG] osdmap e21: 4 total, 3 up, 4 in 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:07.612694+0000 mon.a (mon.0) 418 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:07.612694+0000 mon.a (mon.0) 418 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 2} : dispatch 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:07.612861+0000 mon.a (mon.0) 419 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:07.612861+0000 mon.a (mon.0) 419 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:07.623657+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:07.623657+0000 mon.a (mon.0) 420 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:07.629019+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:07.987 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:07 vm04 bash[20194]: audit 2026-03-21T06:51:07.629019+0000 mon.a (mon.0) 421 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:08.980 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: cluster 2026-03-21T06:51:05.687641+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: cluster 2026-03-21T06:51:05.687641+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: cluster 2026-03-21T06:51:05.687653+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: cluster 2026-03-21T06:51:05.687653+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: cluster 2026-03-21T06:51:07.558560+0000 mgr.x (mgr.14152) 159 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: cluster 2026-03-21T06:51:07.558560+0000 mgr.x (mgr.14152) 159 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:07.775729+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:07.775729+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:08.565294+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:08.565294+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:08.569816+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:08.569816+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:08.630366+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:08.630366+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: cluster 2026-03-21T06:51:08.632722+0000 mon.a (mon.0) 426 : cluster [DBG] osdmap e22: 4 total, 3 up, 4 in 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: cluster 2026-03-21T06:51:08.632722+0000 mon.a (mon.0) 426 : cluster [DBG] osdmap e22: 4 total, 3 up, 4 in 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:08.632992+0000 mon.a (mon.0) 427 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:08.632992+0000 mon.a (mon.0) 427 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:08.633533+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:08.981 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:08 vm04 bash[20194]: audit 2026-03-21T06:51:08.633533+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: cluster 2026-03-21T06:51:05.687641+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: cluster 2026-03-21T06:51:05.687641+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: cluster 2026-03-21T06:51:05.687653+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: cluster 2026-03-21T06:51:05.687653+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: cluster 2026-03-21T06:51:07.558560+0000 mgr.x (mgr.14152) 159 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: cluster 2026-03-21T06:51:07.558560+0000 mgr.x (mgr.14152) 159 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:07.775729+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:07.775729+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:08.565294+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:08.565294+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:08.569816+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:08.569816+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:08.630366+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:08.630366+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: cluster 2026-03-21T06:51:08.632722+0000 mon.a (mon.0) 426 : cluster [DBG] osdmap e22: 4 total, 3 up, 4 in 2026-03-21T06:51:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: cluster 2026-03-21T06:51:08.632722+0000 mon.a (mon.0) 426 : cluster [DBG] osdmap e22: 4 total, 3 up, 4 in 2026-03-21T06:51:09.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:08.632992+0000 mon.a (mon.0) 427 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:09.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:08.632992+0000 mon.a (mon.0) 427 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:09.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:08.633533+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:09.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:08 vm02 bash[17657]: audit 2026-03-21T06:51:08.633533+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: cluster 2026-03-21T06:51:05.687641+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: cluster 2026-03-21T06:51:05.687641+0000 osd.2 (osd.2) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: cluster 2026-03-21T06:51:05.687653+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: cluster 2026-03-21T06:51:05.687653+0000 osd.2 (osd.2) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: cluster 2026-03-21T06:51:07.558560+0000 mgr.x (mgr.14152) 159 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: cluster 2026-03-21T06:51:07.558560+0000 mgr.x (mgr.14152) 159 : cluster [DBG] pgmap v99: 0 pgs: ; 0 B data, 54 MiB used, 40 GiB / 40 GiB avail 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:07.775729+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:07.775729+0000 mon.a (mon.0) 422 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:08.565294+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:08.565294+0000 mon.a (mon.0) 423 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:08.569816+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:08.569816+0000 mon.a (mon.0) 424 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:08.630366+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:08.630366+0000 mon.a (mon.0) 425 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool create", "format": "json", "pool": ".mgr", "pg_num": 1, "pg_num_min": 1, "pg_num_max": 32, "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: cluster 2026-03-21T06:51:08.632722+0000 mon.a (mon.0) 426 : cluster [DBG] osdmap e22: 4 total, 3 up, 4 in 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: cluster 2026-03-21T06:51:08.632722+0000 mon.a (mon.0) 426 : cluster [DBG] osdmap e22: 4 total, 3 up, 4 in 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:08.632992+0000 mon.a (mon.0) 427 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:08.632992+0000 mon.a (mon.0) 427 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:08.633533+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:08 vm07 bash[19945]: audit 2026-03-21T06:51:08.633533+0000 mon.a (mon.0) 428 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true} : dispatch 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.720030+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.720030+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.725641+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.725641+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.726561+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.726561+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.727586+0000 mon.a (mon.0) 432 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.727586+0000 mon.a (mon.0) 432 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cephadm 2026-03-21T06:51:08.728272+0000 mgr.x (mgr.14152) 160 : cephadm [INF] Adjusting osd_memory_target on vm02 to 919.4M 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cephadm 2026-03-21T06:51:08.728272+0000 mgr.x (mgr.14152) 160 : cephadm [INF] Adjusting osd_memory_target on vm02 to 919.4M 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.732250+0000 mon.a (mon.0) 433 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.732250+0000 mon.a (mon.0) 433 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.736739+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.3"} : dispatch 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.736739+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.3"} : dispatch 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.737624+0000 mon.a (mon.0) 435 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:08.737624+0000 mon.a (mon.0) 435 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cephadm 2026-03-21T06:51:08.738339+0000 mgr.x (mgr.14152) 161 : cephadm [INF] Deploying daemon osd.3 on vm04 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cephadm 2026-03-21T06:51:08.738339+0000 mgr.x (mgr.14152) 161 : cephadm [INF] Deploying daemon osd.3 on vm04 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cluster 2026-03-21T06:51:09.558780+0000 mgr.x (mgr.14152) 162 : cluster [DBG] pgmap v102: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cluster 2026-03-21T06:51:09.558780+0000 mgr.x (mgr.14152) 162 : cluster [DBG] pgmap v102: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.633464+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.633464+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cluster 2026-03-21T06:51:09.635602+0000 mon.a (mon.0) 437 : cluster [DBG] osdmap e23: 4 total, 3 up, 4 in 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cluster 2026-03-21T06:51:09.635602+0000 mon.a (mon.0) 437 : cluster [DBG] osdmap e23: 4 total, 3 up, 4 in 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.636789+0000 mon.a (mon.0) 438 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.636789+0000 mon.a (mon.0) 438 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cephadm 2026-03-21T06:51:09.643276+0000 mgr.x (mgr.14152) 163 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cephadm 2026-03-21T06:51:09.643276+0000 mgr.x (mgr.14152) 163 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.648151+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.648151+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.652425+0000 mon.a (mon.0) 440 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.652425+0000 mon.a (mon.0) 440 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.656848+0000 mon.a (mon.0) 441 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.887 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.656848+0000 mon.a (mon.0) 441 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.661390+0000 mon.a (mon.0) 442 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.661390+0000 mon.a (mon.0) 442 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.662030+0000 mon.a (mon.0) 443 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.662030+0000 mon.a (mon.0) 443 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cephadm 2026-03-21T06:51:09.662380+0000 mgr.x (mgr.14152) 164 : cephadm [INF] Adjusting osd_memory_target on vm04 to 5934M 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: cephadm 2026-03-21T06:51:09.662380+0000 mgr.x (mgr.14152) 164 : cephadm [INF] Adjusting osd_memory_target on vm04 to 5934M 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.665082+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.665082+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.666253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.666253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.666709+0000 mon.a (mon.0) 446 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.666709+0000 mon.a (mon.0) 446 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.670558+0000 mon.a (mon.0) 447 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.670558+0000 mon.a (mon.0) 447 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.671444+0000 mon.a (mon.0) 448 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.671444+0000 mon.a (mon.0) 448 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.672865+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.672865+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.673290+0000 mon.a (mon.0) 450 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:09.888 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:09 vm04 bash[20194]: audit 2026-03-21T06:51:09.673290+0000 mon.a (mon.0) 450 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.720030+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.720030+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.725641+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.725641+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.726561+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.726561+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.727586+0000 mon.a (mon.0) 432 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.727586+0000 mon.a (mon.0) 432 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cephadm 2026-03-21T06:51:08.728272+0000 mgr.x (mgr.14152) 160 : cephadm [INF] Adjusting osd_memory_target on vm02 to 919.4M 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cephadm 2026-03-21T06:51:08.728272+0000 mgr.x (mgr.14152) 160 : cephadm [INF] Adjusting osd_memory_target on vm02 to 919.4M 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.732250+0000 mon.a (mon.0) 433 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.732250+0000 mon.a (mon.0) 433 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.736739+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.3"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.736739+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.3"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.737624+0000 mon.a (mon.0) 435 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:08.737624+0000 mon.a (mon.0) 435 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cephadm 2026-03-21T06:51:08.738339+0000 mgr.x (mgr.14152) 161 : cephadm [INF] Deploying daemon osd.3 on vm04 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cephadm 2026-03-21T06:51:08.738339+0000 mgr.x (mgr.14152) 161 : cephadm [INF] Deploying daemon osd.3 on vm04 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cluster 2026-03-21T06:51:09.558780+0000 mgr.x (mgr.14152) 162 : cluster [DBG] pgmap v102: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cluster 2026-03-21T06:51:09.558780+0000 mgr.x (mgr.14152) 162 : cluster [DBG] pgmap v102: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.633464+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.633464+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cluster 2026-03-21T06:51:09.635602+0000 mon.a (mon.0) 437 : cluster [DBG] osdmap e23: 4 total, 3 up, 4 in 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cluster 2026-03-21T06:51:09.635602+0000 mon.a (mon.0) 437 : cluster [DBG] osdmap e23: 4 total, 3 up, 4 in 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.636789+0000 mon.a (mon.0) 438 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.636789+0000 mon.a (mon.0) 438 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cephadm 2026-03-21T06:51:09.643276+0000 mgr.x (mgr.14152) 163 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cephadm 2026-03-21T06:51:09.643276+0000 mgr.x (mgr.14152) 163 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.648151+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.648151+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.652425+0000 mon.a (mon.0) 440 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.652425+0000 mon.a (mon.0) 440 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.656848+0000 mon.a (mon.0) 441 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.656848+0000 mon.a (mon.0) 441 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.661390+0000 mon.a (mon.0) 442 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.661390+0000 mon.a (mon.0) 442 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.662030+0000 mon.a (mon.0) 443 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.662030+0000 mon.a (mon.0) 443 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cephadm 2026-03-21T06:51:09.662380+0000 mgr.x (mgr.14152) 164 : cephadm [INF] Adjusting osd_memory_target on vm04 to 5934M 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: cephadm 2026-03-21T06:51:09.662380+0000 mgr.x (mgr.14152) 164 : cephadm [INF] Adjusting osd_memory_target on vm04 to 5934M 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.665082+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.665082+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.666253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.666253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.666709+0000 mon.a (mon.0) 446 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.666709+0000 mon.a (mon.0) 446 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.670558+0000 mon.a (mon.0) 447 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.670558+0000 mon.a (mon.0) 447 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.671444+0000 mon.a (mon.0) 448 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.671444+0000 mon.a (mon.0) 448 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.672865+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.672865+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.673290+0000 mon.a (mon.0) 450 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:09 vm02 bash[17657]: audit 2026-03-21T06:51:09.673290+0000 mon.a (mon.0) 450 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.720030+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.720030+0000 mon.a (mon.0) 429 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.725641+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.725641+0000 mon.a (mon.0) 430 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.726561+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.726561+0000 mon.a (mon.0) 431 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.727586+0000 mon.a (mon.0) 432 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.727586+0000 mon.a (mon.0) 432 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cephadm 2026-03-21T06:51:08.728272+0000 mgr.x (mgr.14152) 160 : cephadm [INF] Adjusting osd_memory_target on vm02 to 919.4M 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cephadm 2026-03-21T06:51:08.728272+0000 mgr.x (mgr.14152) 160 : cephadm [INF] Adjusting osd_memory_target on vm02 to 919.4M 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.732250+0000 mon.a (mon.0) 433 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.732250+0000 mon.a (mon.0) 433 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.736739+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.3"} : dispatch 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.736739+0000 mon.a (mon.0) 434 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.3"} : dispatch 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.737624+0000 mon.a (mon.0) 435 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:08.737624+0000 mon.a (mon.0) 435 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cephadm 2026-03-21T06:51:08.738339+0000 mgr.x (mgr.14152) 161 : cephadm [INF] Deploying daemon osd.3 on vm04 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cephadm 2026-03-21T06:51:08.738339+0000 mgr.x (mgr.14152) 161 : cephadm [INF] Deploying daemon osd.3 on vm04 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cluster 2026-03-21T06:51:09.558780+0000 mgr.x (mgr.14152) 162 : cluster [DBG] pgmap v102: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cluster 2026-03-21T06:51:09.558780+0000 mgr.x (mgr.14152) 162 : cluster [DBG] pgmap v102: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.633464+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.633464+0000 mon.a (mon.0) 436 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "osd pool application enable", "format": "json", "pool": ".mgr", "app": "mgr", "yes_i_really_mean_it": true}]': finished 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cluster 2026-03-21T06:51:09.635602+0000 mon.a (mon.0) 437 : cluster [DBG] osdmap e23: 4 total, 3 up, 4 in 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cluster 2026-03-21T06:51:09.635602+0000 mon.a (mon.0) 437 : cluster [DBG] osdmap e23: 4 total, 3 up, 4 in 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.636789+0000 mon.a (mon.0) 438 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.636789+0000 mon.a (mon.0) 438 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cephadm 2026-03-21T06:51:09.643276+0000 mgr.x (mgr.14152) 163 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cephadm 2026-03-21T06:51:09.643276+0000 mgr.x (mgr.14152) 163 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.648151+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.648151+0000 mon.a (mon.0) 439 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.652425+0000 mon.a (mon.0) 440 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.652425+0000 mon.a (mon.0) 440 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.656848+0000 mon.a (mon.0) 441 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.656848+0000 mon.a (mon.0) 441 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.661390+0000 mon.a (mon.0) 442 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.661390+0000 mon.a (mon.0) 442 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.662030+0000 mon.a (mon.0) 443 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.662030+0000 mon.a (mon.0) 443 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cephadm 2026-03-21T06:51:09.662380+0000 mgr.x (mgr.14152) 164 : cephadm [INF] Adjusting osd_memory_target on vm04 to 5934M 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: cephadm 2026-03-21T06:51:09.662380+0000 mgr.x (mgr.14152) 164 : cephadm [INF] Adjusting osd_memory_target on vm04 to 5934M 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.665082+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.665082+0000 mon.a (mon.0) 444 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.666253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.666253+0000 mon.a (mon.0) 445 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.666709+0000 mon.a (mon.0) 446 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.666709+0000 mon.a (mon.0) 446 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.670558+0000 mon.a (mon.0) 447 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.670558+0000 mon.a (mon.0) 447 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.671444+0000 mon.a (mon.0) 448 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.671444+0000 mon.a (mon.0) 448 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.672865+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.672865+0000 mon.a (mon.0) 449 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.673290+0000 mon.a (mon.0) 450 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:09 vm07 bash[19945]: audit 2026-03-21T06:51:09.673290+0000 mon.a (mon.0) 450 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:10.490 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:10.490 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:10.490 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:51:10 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:10.490 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:51:10 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:11.061 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.442318+0000 mon.a (mon.0) 451 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.442318+0000 mon.a (mon.0) 451 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.461599+0000 mon.a (mon.0) 452 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.461599+0000 mon.a (mon.0) 452 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.461836+0000 mon.a (mon.0) 453 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.461836+0000 mon.a (mon.0) 453 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.461925+0000 mon.a (mon.0) 454 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.461925+0000 mon.a (mon.0) 454 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.461989+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.461989+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.463592+0000 mon.a (mon.0) 456 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.463592+0000 mon.a (mon.0) 456 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.463658+0000 mon.a (mon.0) 457 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.463658+0000 mon.a (mon.0) 457 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.463718+0000 mon.a (mon.0) 458 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.463718+0000 mon.a (mon.0) 458 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.468669+0000 mon.c (mon.1) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.468669+0000 mon.c (mon.1) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.486957+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.486957+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.487087+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.487087+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.487243+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.487243+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.489301+0000 mon.c (mon.1) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.489301+0000 mon.c (mon.1) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.495470+0000 mon.b (mon.2) 10 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.495470+0000 mon.b (mon.2) 10 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.514020+0000 mon.a (mon.0) 462 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.514020+0000 mon.a (mon.0) 462 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.515295+0000 mon.b (mon.2) 11 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.515295+0000 mon.b (mon.2) 11 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.519086+0000 mon.a (mon.0) 463 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:11.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:10 vm04 bash[20194]: audit 2026-03-21T06:51:10.519086+0000 mon.a (mon.0) 463 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.442318+0000 mon.a (mon.0) 451 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.442318+0000 mon.a (mon.0) 451 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.461599+0000 mon.a (mon.0) 452 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.461599+0000 mon.a (mon.0) 452 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.461836+0000 mon.a (mon.0) 453 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.461836+0000 mon.a (mon.0) 453 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.461925+0000 mon.a (mon.0) 454 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.461925+0000 mon.a (mon.0) 454 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.461989+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.461989+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.463592+0000 mon.a (mon.0) 456 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.463592+0000 mon.a (mon.0) 456 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.463658+0000 mon.a (mon.0) 457 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.463658+0000 mon.a (mon.0) 457 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.463718+0000 mon.a (mon.0) 458 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.463718+0000 mon.a (mon.0) 458 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.468669+0000 mon.c (mon.1) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.468669+0000 mon.c (mon.1) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.486957+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.486957+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.487087+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.487087+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.487243+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.487243+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.489301+0000 mon.c (mon.1) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.489301+0000 mon.c (mon.1) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.495470+0000 mon.b (mon.2) 10 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.495470+0000 mon.b (mon.2) 10 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.514020+0000 mon.a (mon.0) 462 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.514020+0000 mon.a (mon.0) 462 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:11.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.515295+0000 mon.b (mon.2) 11 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.515295+0000 mon.b (mon.2) 11 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.519086+0000 mon.a (mon.0) 463 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:11.147 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:10 vm02 bash[17657]: audit 2026-03-21T06:51:10.519086+0000 mon.a (mon.0) 463 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.442318+0000 mon.a (mon.0) 451 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.442318+0000 mon.a (mon.0) 451 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.461599+0000 mon.a (mon.0) 452 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.461599+0000 mon.a (mon.0) 452 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.461836+0000 mon.a (mon.0) 453 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.461836+0000 mon.a (mon.0) 453 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.461925+0000 mon.a (mon.0) 454 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.461925+0000 mon.a (mon.0) 454 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.461989+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.461989+0000 mon.a (mon.0) 455 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.463592+0000 mon.a (mon.0) 456 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.463592+0000 mon.a (mon.0) 456 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.463658+0000 mon.a (mon.0) 457 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.463658+0000 mon.a (mon.0) 457 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.463718+0000 mon.a (mon.0) 458 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.463718+0000 mon.a (mon.0) 458 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.468669+0000 mon.c (mon.1) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.468669+0000 mon.c (mon.1) 6 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.486957+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.486957+0000 mon.a (mon.0) 459 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "a"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.487087+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.487087+0000 mon.a (mon.0) 460 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "b"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.487243+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.487243+0000 mon.a (mon.0) 461 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "mon metadata", "id": "c"} : dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.489301+0000 mon.c (mon.1) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.489301+0000 mon.c (mon.1) 7 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.495470+0000 mon.b (mon.2) 10 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.495470+0000 mon.b (mon.2) 10 : audit [INF] from='admin socket' entity='admin socket' cmd='smart' args=[json]: dispatch 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.514020+0000 mon.a (mon.0) 462 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.514020+0000 mon.a (mon.0) 462 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.515295+0000 mon.b (mon.2) 11 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.515295+0000 mon.b (mon.2) 11 : audit [INF] from='admin socket' entity='admin socket' cmd=smart args=[json]: finished 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.519086+0000 mon.a (mon.0) 463 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:10 vm07 bash[19945]: audit 2026-03-21T06:51:10.519086+0000 mon.a (mon.0) 463 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:12.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:11 vm04 bash[20194]: cluster 2026-03-21T06:51:10.744558+0000 mon.a (mon.0) 464 : cluster [DBG] osdmap e24: 4 total, 3 up, 4 in 2026-03-21T06:51:12.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:11 vm04 bash[20194]: cluster 2026-03-21T06:51:10.744558+0000 mon.a (mon.0) 464 : cluster [DBG] osdmap e24: 4 total, 3 up, 4 in 2026-03-21T06:51:12.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:11 vm04 bash[20194]: audit 2026-03-21T06:51:10.745261+0000 mon.a (mon.0) 465 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:12.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:11 vm04 bash[20194]: audit 2026-03-21T06:51:10.745261+0000 mon.a (mon.0) 465 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:12.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:11 vm04 bash[20194]: cluster 2026-03-21T06:51:11.559024+0000 mgr.x (mgr.14152) 165 : cluster [DBG] pgmap v105: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:12.062 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:11 vm04 bash[20194]: cluster 2026-03-21T06:51:11.559024+0000 mgr.x (mgr.14152) 165 : cluster [DBG] pgmap v105: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:12.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:11 vm02 bash[17657]: cluster 2026-03-21T06:51:10.744558+0000 mon.a (mon.0) 464 : cluster [DBG] osdmap e24: 4 total, 3 up, 4 in 2026-03-21T06:51:12.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:11 vm02 bash[17657]: cluster 2026-03-21T06:51:10.744558+0000 mon.a (mon.0) 464 : cluster [DBG] osdmap e24: 4 total, 3 up, 4 in 2026-03-21T06:51:12.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:11 vm02 bash[17657]: audit 2026-03-21T06:51:10.745261+0000 mon.a (mon.0) 465 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:12.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:11 vm02 bash[17657]: audit 2026-03-21T06:51:10.745261+0000 mon.a (mon.0) 465 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:12.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:11 vm02 bash[17657]: cluster 2026-03-21T06:51:11.559024+0000 mgr.x (mgr.14152) 165 : cluster [DBG] pgmap v105: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:12.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:11 vm02 bash[17657]: cluster 2026-03-21T06:51:11.559024+0000 mgr.x (mgr.14152) 165 : cluster [DBG] pgmap v105: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:12.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:11 vm07 bash[19945]: cluster 2026-03-21T06:51:10.744558+0000 mon.a (mon.0) 464 : cluster [DBG] osdmap e24: 4 total, 3 up, 4 in 2026-03-21T06:51:12.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:11 vm07 bash[19945]: cluster 2026-03-21T06:51:10.744558+0000 mon.a (mon.0) 464 : cluster [DBG] osdmap e24: 4 total, 3 up, 4 in 2026-03-21T06:51:12.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:11 vm07 bash[19945]: audit 2026-03-21T06:51:10.745261+0000 mon.a (mon.0) 465 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:12.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:11 vm07 bash[19945]: audit 2026-03-21T06:51:10.745261+0000 mon.a (mon.0) 465 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:12.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:11 vm07 bash[19945]: cluster 2026-03-21T06:51:11.559024+0000 mgr.x (mgr.14152) 165 : cluster [DBG] pgmap v105: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:12.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:11 vm07 bash[19945]: cluster 2026-03-21T06:51:11.559024+0000 mgr.x (mgr.14152) 165 : cluster [DBG] pgmap v105: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:12.284 INFO:teuthology.orchestra.run.vm04.stdout:Created osd(s) 3 on host 'vm04' 2026-03-21T06:51:12.349 DEBUG:teuthology.orchestra.run.vm04:osd.3> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.3.service 2026-03-21T06:51:12.351 INFO:tasks.cephadm:Deploying osd.4 on vm04 with /dev/vg_nvme/lv_2... 2026-03-21T06:51:12.351 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- lvm zap /dev/vg_nvme/lv_2 2026-03-21T06:51:12.643 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.b/config 2026-03-21T06:51:12.679 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:51:12 vm04 bash[29705]: debug 2026-03-21T06:51:12.487+0000 7f38cdb678c0 -1 Falling back to public interface 2026-03-21T06:51:12.933 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:12 vm04 bash[20194]: cluster 2026-03-21T06:51:11.786691+0000 mon.a (mon.0) 466 : cluster [DBG] mgrmap e14: x(active, since 3m) 2026-03-21T06:51:12.933 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:12 vm04 bash[20194]: cluster 2026-03-21T06:51:11.786691+0000 mon.a (mon.0) 466 : cluster [DBG] mgrmap e14: x(active, since 3m) 2026-03-21T06:51:12.933 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:12 vm04 bash[20194]: audit 2026-03-21T06:51:12.271757+0000 mon.a (mon.0) 467 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:12.933 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:12 vm04 bash[20194]: audit 2026-03-21T06:51:12.271757+0000 mon.a (mon.0) 467 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:12.933 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:12 vm04 bash[20194]: audit 2026-03-21T06:51:12.275920+0000 mon.a (mon.0) 468 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:12.933 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:12 vm04 bash[20194]: audit 2026-03-21T06:51:12.275920+0000 mon.a (mon.0) 468 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:12.933 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:12 vm04 bash[20194]: audit 2026-03-21T06:51:12.780521+0000 mon.a (mon.0) 469 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:12.934 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:12 vm04 bash[20194]: audit 2026-03-21T06:51:12.780521+0000 mon.a (mon.0) 469 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:12.934 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:12 vm04 bash[20194]: audit 2026-03-21T06:51:12.788716+0000 mon.b (mon.2) 12 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:12.934 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:12 vm04 bash[20194]: audit 2026-03-21T06:51:12.788716+0000 mon.b (mon.2) 12 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:12.934 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:51:12 vm04 bash[29705]: debug 2026-03-21T06:51:12.775+0000 7f38cdb678c0 -1 osd.3 0 log_to_monitors true 2026-03-21T06:51:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:12 vm02 bash[17657]: cluster 2026-03-21T06:51:11.786691+0000 mon.a (mon.0) 466 : cluster [DBG] mgrmap e14: x(active, since 3m) 2026-03-21T06:51:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:12 vm02 bash[17657]: cluster 2026-03-21T06:51:11.786691+0000 mon.a (mon.0) 466 : cluster [DBG] mgrmap e14: x(active, since 3m) 2026-03-21T06:51:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:12 vm02 bash[17657]: audit 2026-03-21T06:51:12.271757+0000 mon.a (mon.0) 467 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:12 vm02 bash[17657]: audit 2026-03-21T06:51:12.271757+0000 mon.a (mon.0) 467 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:12 vm02 bash[17657]: audit 2026-03-21T06:51:12.275920+0000 mon.a (mon.0) 468 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:12 vm02 bash[17657]: audit 2026-03-21T06:51:12.275920+0000 mon.a (mon.0) 468 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:12 vm02 bash[17657]: audit 2026-03-21T06:51:12.780521+0000 mon.a (mon.0) 469 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:12 vm02 bash[17657]: audit 2026-03-21T06:51:12.780521+0000 mon.a (mon.0) 469 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:12 vm02 bash[17657]: audit 2026-03-21T06:51:12.788716+0000 mon.b (mon.2) 12 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:13.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:12 vm02 bash[17657]: audit 2026-03-21T06:51:12.788716+0000 mon.b (mon.2) 12 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:12 vm07 bash[19945]: cluster 2026-03-21T06:51:11.786691+0000 mon.a (mon.0) 466 : cluster [DBG] mgrmap e14: x(active, since 3m) 2026-03-21T06:51:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:12 vm07 bash[19945]: cluster 2026-03-21T06:51:11.786691+0000 mon.a (mon.0) 466 : cluster [DBG] mgrmap e14: x(active, since 3m) 2026-03-21T06:51:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:12 vm07 bash[19945]: audit 2026-03-21T06:51:12.271757+0000 mon.a (mon.0) 467 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:12 vm07 bash[19945]: audit 2026-03-21T06:51:12.271757+0000 mon.a (mon.0) 467 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:12 vm07 bash[19945]: audit 2026-03-21T06:51:12.275920+0000 mon.a (mon.0) 468 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:12 vm07 bash[19945]: audit 2026-03-21T06:51:12.275920+0000 mon.a (mon.0) 468 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:12 vm07 bash[19945]: audit 2026-03-21T06:51:12.780521+0000 mon.a (mon.0) 469 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:12 vm07 bash[19945]: audit 2026-03-21T06:51:12.780521+0000 mon.a (mon.0) 469 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:12 vm07 bash[19945]: audit 2026-03-21T06:51:12.788716+0000 mon.b (mon.2) 12 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:12 vm07 bash[19945]: audit 2026-03-21T06:51:12.788716+0000 mon.b (mon.2) 12 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]} : dispatch 2026-03-21T06:51:13.212 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:51:13.225 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch daemon add osd vm04:vg_nvme/lv_2 --skip-validation 2026-03-21T06:51:13.468 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.b/config 2026-03-21T06:51:14.460 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.278750+0000 mon.a (mon.0) 470 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-21T06:51:14.460 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.278750+0000 mon.a (mon.0) 470 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: cluster 2026-03-21T06:51:13.281390+0000 mon.a (mon.0) 471 : cluster [DBG] osdmap e25: 4 total, 3 up, 4 in 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: cluster 2026-03-21T06:51:13.281390+0000 mon.a (mon.0) 471 : cluster [DBG] osdmap e25: 4 total, 3 up, 4 in 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.281630+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.281630+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.282774+0000 mon.a (mon.0) 473 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.282774+0000 mon.a (mon.0) 473 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.291046+0000 mon.b (mon.2) 13 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.291046+0000 mon.b (mon.2) 13 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: cluster 2026-03-21T06:51:13.559278+0000 mgr.x (mgr.14152) 166 : cluster [DBG] pgmap v107: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: cluster 2026-03-21T06:51:13.559278+0000 mgr.x (mgr.14152) 166 : cluster [DBG] pgmap v107: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.859438+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.859438+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.863369+0000 mon.a (mon.0) 475 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.863369+0000 mon.a (mon.0) 475 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.864339+0000 mon.a (mon.0) 476 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.864339+0000 mon.a (mon.0) 476 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.865786+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.865786+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.866265+0000 mon.a (mon.0) 478 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:14 vm04 bash[20194]: audit 2026-03-21T06:51:13.866265+0000 mon.a (mon.0) 478 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:14.461 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:51:14 vm04 bash[29705]: debug 2026-03-21T06:51:14.311+0000 7f38c9ae7640 -1 osd.3 0 waiting for initial osdmap 2026-03-21T06:51:14.461 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:51:14 vm04 bash[29705]: debug 2026-03-21T06:51:14.315+0000 7f38c50cf640 -1 osd.3 26 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.278750+0000 mon.a (mon.0) 470 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.278750+0000 mon.a (mon.0) 470 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: cluster 2026-03-21T06:51:13.281390+0000 mon.a (mon.0) 471 : cluster [DBG] osdmap e25: 4 total, 3 up, 4 in 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: cluster 2026-03-21T06:51:13.281390+0000 mon.a (mon.0) 471 : cluster [DBG] osdmap e25: 4 total, 3 up, 4 in 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.281630+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.281630+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.282774+0000 mon.a (mon.0) 473 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.282774+0000 mon.a (mon.0) 473 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.291046+0000 mon.b (mon.2) 13 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.291046+0000 mon.b (mon.2) 13 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: cluster 2026-03-21T06:51:13.559278+0000 mgr.x (mgr.14152) 166 : cluster [DBG] pgmap v107: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: cluster 2026-03-21T06:51:13.559278+0000 mgr.x (mgr.14152) 166 : cluster [DBG] pgmap v107: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.859438+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.859438+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.863369+0000 mon.a (mon.0) 475 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.863369+0000 mon.a (mon.0) 475 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.864339+0000 mon.a (mon.0) 476 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.864339+0000 mon.a (mon.0) 476 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.865786+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.865786+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.866265+0000 mon.a (mon.0) 478 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:14.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:14 vm02 bash[17657]: audit 2026-03-21T06:51:13.866265+0000 mon.a (mon.0) 478 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.278750+0000 mon.a (mon.0) 470 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.278750+0000 mon.a (mon.0) 470 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["3"]}]': finished 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: cluster 2026-03-21T06:51:13.281390+0000 mon.a (mon.0) 471 : cluster [DBG] osdmap e25: 4 total, 3 up, 4 in 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: cluster 2026-03-21T06:51:13.281390+0000 mon.a (mon.0) 471 : cluster [DBG] osdmap e25: 4 total, 3 up, 4 in 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.281630+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.281630+0000 mon.a (mon.0) 472 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.282774+0000 mon.a (mon.0) 473 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.282774+0000 mon.a (mon.0) 473 : audit [INF] from='osd.3 ' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.291046+0000 mon.b (mon.2) 13 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.291046+0000 mon.b (mon.2) 13 : audit [INF] from='osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205]' entity='osd.3' cmd={"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: cluster 2026-03-21T06:51:13.559278+0000 mgr.x (mgr.14152) 166 : cluster [DBG] pgmap v107: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: cluster 2026-03-21T06:51:13.559278+0000 mgr.x (mgr.14152) 166 : cluster [DBG] pgmap v107: 1 pgs: 1 unknown; 0 B data, 80 MiB used, 60 GiB / 60 GiB avail 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.859438+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.859438+0000 mon.a (mon.0) 474 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.863369+0000 mon.a (mon.0) 475 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.863369+0000 mon.a (mon.0) 475 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:14.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.864339+0000 mon.a (mon.0) 476 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:14.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.864339+0000 mon.a (mon.0) 476 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:14.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.865786+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:14.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.865786+0000 mon.a (mon.0) 477 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:14.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.866265+0000 mon.a (mon.0) 478 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:14.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:14 vm07 bash[19945]: audit 2026-03-21T06:51:13.866265+0000 mon.a (mon.0) 478 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:13.853831+0000 mgr.x (mgr.14152) 167 : audit [DBG] from='client.24233 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:13.853831+0000 mgr.x (mgr.14152) 167 : audit [DBG] from='client.24233 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cephadm 2026-03-21T06:51:13.854693+0000 mgr.x (mgr.14152) 168 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cephadm 2026-03-21T06:51:13.854693+0000 mgr.x (mgr.14152) 168 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cephadm 2026-03-21T06:51:13.854761+0000 mgr.x (mgr.14152) 169 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_2'] 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cephadm 2026-03-21T06:51:13.854761+0000 mgr.x (mgr.14152) 169 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_2'] 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cephadm 2026-03-21T06:51:13.859698+0000 mgr.x (mgr.14152) 170 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cephadm 2026-03-21T06:51:13.859698+0000 mgr.x (mgr.14152) 170 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cephadm 2026-03-21T06:51:13.859795+0000 mgr.x (mgr.14152) 171 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cephadm 2026-03-21T06:51:13.859795+0000 mgr.x (mgr.14152) 171 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.284905+0000 mon.a (mon.0) 479 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.284905+0000 mon.a (mon.0) 479 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cluster 2026-03-21T06:51:14.290768+0000 mon.a (mon.0) 480 : cluster [DBG] osdmap e26: 4 total, 3 up, 4 in 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cluster 2026-03-21T06:51:14.290768+0000 mon.a (mon.0) 480 : cluster [DBG] osdmap e26: 4 total, 3 up, 4 in 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.291108+0000 mon.a (mon.0) 481 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.291108+0000 mon.a (mon.0) 481 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.291936+0000 mon.a (mon.0) 482 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.291936+0000 mon.a (mon.0) 482 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.522580+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.522580+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.526575+0000 mon.a (mon.0) 484 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.526575+0000 mon.a (mon.0) 484 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.542038+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:14.542038+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:15.070731+0000 mon.a (mon.0) 486 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:15.070731+0000 mon.a (mon.0) 486 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:15.074979+0000 mon.a (mon.0) 487 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"}]': finished 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:15.074979+0000 mon.a (mon.0) 487 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"}]': finished 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cluster 2026-03-21T06:51:15.078214+0000 mon.a (mon.0) 488 : cluster [INF] osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205] boot 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cluster 2026-03-21T06:51:15.078214+0000 mon.a (mon.0) 488 : cluster [INF] osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205] boot 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cluster 2026-03-21T06:51:15.078286+0000 mon.a (mon.0) 489 : cluster [DBG] osdmap e27: 5 total, 4 up, 5 in 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: cluster 2026-03-21T06:51:15.078286+0000 mon.a (mon.0) 489 : cluster [DBG] osdmap e27: 5 total, 4 up, 5 in 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:15.078681+0000 mon.a (mon.0) 490 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:15.078681+0000 mon.a (mon.0) 490 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:15.078840+0000 mon.a (mon.0) 491 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:15.078840+0000 mon.a (mon.0) 491 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:15.078924+0000 mon.b (mon.2) 14 : audit [INF] from='client.? 192.168.123.104:0/42804440' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:15.647 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:15 vm02 bash[17657]: audit 2026-03-21T06:51:15.078924+0000 mon.b (mon.2) 14 : audit [INF] from='client.? 192.168.123.104:0/42804440' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:15.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:13.853831+0000 mgr.x (mgr.14152) 167 : audit [DBG] from='client.24233 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:15.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:13.853831+0000 mgr.x (mgr.14152) 167 : audit [DBG] from='client.24233 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:15.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cephadm 2026-03-21T06:51:13.854693+0000 mgr.x (mgr.14152) 168 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:15.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cephadm 2026-03-21T06:51:13.854693+0000 mgr.x (mgr.14152) 168 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cephadm 2026-03-21T06:51:13.854761+0000 mgr.x (mgr.14152) 169 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_2'] 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cephadm 2026-03-21T06:51:13.854761+0000 mgr.x (mgr.14152) 169 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_2'] 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cephadm 2026-03-21T06:51:13.859698+0000 mgr.x (mgr.14152) 170 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cephadm 2026-03-21T06:51:13.859698+0000 mgr.x (mgr.14152) 170 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cephadm 2026-03-21T06:51:13.859795+0000 mgr.x (mgr.14152) 171 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cephadm 2026-03-21T06:51:13.859795+0000 mgr.x (mgr.14152) 171 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.284905+0000 mon.a (mon.0) 479 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.284905+0000 mon.a (mon.0) 479 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cluster 2026-03-21T06:51:14.290768+0000 mon.a (mon.0) 480 : cluster [DBG] osdmap e26: 4 total, 3 up, 4 in 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cluster 2026-03-21T06:51:14.290768+0000 mon.a (mon.0) 480 : cluster [DBG] osdmap e26: 4 total, 3 up, 4 in 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.291108+0000 mon.a (mon.0) 481 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.291108+0000 mon.a (mon.0) 481 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.291936+0000 mon.a (mon.0) 482 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.291936+0000 mon.a (mon.0) 482 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.522580+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.522580+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.526575+0000 mon.a (mon.0) 484 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.526575+0000 mon.a (mon.0) 484 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.542038+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:14.542038+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:15.070731+0000 mon.a (mon.0) 486 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:15.070731+0000 mon.a (mon.0) 486 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:15.074979+0000 mon.a (mon.0) 487 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"}]': finished 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:15.074979+0000 mon.a (mon.0) 487 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"}]': finished 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cluster 2026-03-21T06:51:15.078214+0000 mon.a (mon.0) 488 : cluster [INF] osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205] boot 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cluster 2026-03-21T06:51:15.078214+0000 mon.a (mon.0) 488 : cluster [INF] osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205] boot 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cluster 2026-03-21T06:51:15.078286+0000 mon.a (mon.0) 489 : cluster [DBG] osdmap e27: 5 total, 4 up, 5 in 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: cluster 2026-03-21T06:51:15.078286+0000 mon.a (mon.0) 489 : cluster [DBG] osdmap e27: 5 total, 4 up, 5 in 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:15.078681+0000 mon.a (mon.0) 490 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:15.078681+0000 mon.a (mon.0) 490 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:15.078840+0000 mon.a (mon.0) 491 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:15.078840+0000 mon.a (mon.0) 491 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:15.078924+0000 mon.b (mon.2) 14 : audit [INF] from='client.? 192.168.123.104:0/42804440' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:15.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:15 vm07 bash[19945]: audit 2026-03-21T06:51:15.078924+0000 mon.b (mon.2) 14 : audit [INF] from='client.? 192.168.123.104:0/42804440' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:13.853831+0000 mgr.x (mgr.14152) 167 : audit [DBG] from='client.24233 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:13.853831+0000 mgr.x (mgr.14152) 167 : audit [DBG] from='client.24233 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm04:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cephadm 2026-03-21T06:51:13.854693+0000 mgr.x (mgr.14152) 168 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cephadm 2026-03-21T06:51:13.854693+0000 mgr.x (mgr.14152) 168 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cephadm 2026-03-21T06:51:13.854761+0000 mgr.x (mgr.14152) 169 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_2'] 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cephadm 2026-03-21T06:51:13.854761+0000 mgr.x (mgr.14152) 169 : cephadm [INF] Creating OSDs with service ID: default on vm04:['vg_nvme/lv_2'] 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cephadm 2026-03-21T06:51:13.859698+0000 mgr.x (mgr.14152) 170 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cephadm 2026-03-21T06:51:13.859698+0000 mgr.x (mgr.14152) 170 : cephadm [INF] Marking host: vm04 for OSDSpec preview refresh. 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cephadm 2026-03-21T06:51:13.859795+0000 mgr.x (mgr.14152) 171 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cephadm 2026-03-21T06:51:13.859795+0000 mgr.x (mgr.14152) 171 : cephadm [INF] Saving service osd.default spec with placement vm04 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.284905+0000 mon.a (mon.0) 479 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.284905+0000 mon.a (mon.0) 479 : audit [INF] from='osd.3 ' entity='osd.3' cmd='[{"prefix": "osd crush create-or-move", "id": 3, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cluster 2026-03-21T06:51:14.290768+0000 mon.a (mon.0) 480 : cluster [DBG] osdmap e26: 4 total, 3 up, 4 in 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cluster 2026-03-21T06:51:14.290768+0000 mon.a (mon.0) 480 : cluster [DBG] osdmap e26: 4 total, 3 up, 4 in 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.291108+0000 mon.a (mon.0) 481 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.291108+0000 mon.a (mon.0) 481 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.291936+0000 mon.a (mon.0) 482 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.291936+0000 mon.a (mon.0) 482 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.522580+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.522580+0000 mon.a (mon.0) 483 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.526575+0000 mon.a (mon.0) 484 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.526575+0000 mon.a (mon.0) 484 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.542038+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:14.542038+0000 mon.a (mon.0) 485 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:15.070731+0000 mon.a (mon.0) 486 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:15.070731+0000 mon.a (mon.0) 486 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:15.074979+0000 mon.a (mon.0) 487 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"}]': finished 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:15.074979+0000 mon.a (mon.0) 487 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"}]': finished 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cluster 2026-03-21T06:51:15.078214+0000 mon.a (mon.0) 488 : cluster [INF] osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205] boot 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cluster 2026-03-21T06:51:15.078214+0000 mon.a (mon.0) 488 : cluster [INF] osd.3 [v2:192.168.123.104:6808/3282486205,v1:192.168.123.104:6809/3282486205] boot 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cluster 2026-03-21T06:51:15.078286+0000 mon.a (mon.0) 489 : cluster [DBG] osdmap e27: 5 total, 4 up, 5 in 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: cluster 2026-03-21T06:51:15.078286+0000 mon.a (mon.0) 489 : cluster [DBG] osdmap e27: 5 total, 4 up, 5 in 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:15.078681+0000 mon.a (mon.0) 490 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:15.078681+0000 mon.a (mon.0) 490 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 3} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:15.078840+0000 mon.a (mon.0) 491 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:15.078840+0000 mon.a (mon.0) 491 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:15.078924+0000 mon.b (mon.2) 14 : audit [INF] from='client.? 192.168.123.104:0/42804440' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:15.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:15 vm04 bash[20194]: audit 2026-03-21T06:51:15.078924+0000 mon.b (mon.2) 14 : audit [INF] from='client.? 192.168.123.104:0/42804440' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "22bec0cb-fc4e-4173-bab0-1ea7c238683c"} : dispatch 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: cluster 2026-03-21T06:51:13.735234+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: cluster 2026-03-21T06:51:13.735234+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: cluster 2026-03-21T06:51:13.735247+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: cluster 2026-03-21T06:51:13.735247+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: audit 2026-03-21T06:51:15.431180+0000 mon.b (mon.2) 15 : audit [DBG] from='client.? 192.168.123.104:0/3495329696' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: audit 2026-03-21T06:51:15.431180+0000 mon.b (mon.2) 15 : audit [DBG] from='client.? 192.168.123.104:0/3495329696' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: cluster 2026-03-21T06:51:15.559550+0000 mgr.x (mgr.14152) 172 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: cluster 2026-03-21T06:51:15.559550+0000 mgr.x (mgr.14152) 172 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: cluster 2026-03-21T06:51:16.080201+0000 mon.a (mon.0) 492 : cluster [DBG] osdmap e28: 5 total, 4 up, 5 in 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: cluster 2026-03-21T06:51:16.080201+0000 mon.a (mon.0) 492 : cluster [DBG] osdmap e28: 5 total, 4 up, 5 in 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: audit 2026-03-21T06:51:16.080392+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:16.646 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:16 vm02 bash[17657]: audit 2026-03-21T06:51:16.080392+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: cluster 2026-03-21T06:51:13.735234+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: cluster 2026-03-21T06:51:13.735234+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: cluster 2026-03-21T06:51:13.735247+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: cluster 2026-03-21T06:51:13.735247+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: audit 2026-03-21T06:51:15.431180+0000 mon.b (mon.2) 15 : audit [DBG] from='client.? 192.168.123.104:0/3495329696' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: audit 2026-03-21T06:51:15.431180+0000 mon.b (mon.2) 15 : audit [DBG] from='client.? 192.168.123.104:0/3495329696' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: cluster 2026-03-21T06:51:15.559550+0000 mgr.x (mgr.14152) 172 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: cluster 2026-03-21T06:51:15.559550+0000 mgr.x (mgr.14152) 172 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: cluster 2026-03-21T06:51:16.080201+0000 mon.a (mon.0) 492 : cluster [DBG] osdmap e28: 5 total, 4 up, 5 in 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: cluster 2026-03-21T06:51:16.080201+0000 mon.a (mon.0) 492 : cluster [DBG] osdmap e28: 5 total, 4 up, 5 in 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: audit 2026-03-21T06:51:16.080392+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:16.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:16 vm07 bash[19945]: audit 2026-03-21T06:51:16.080392+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: cluster 2026-03-21T06:51:13.735234+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: cluster 2026-03-21T06:51:13.735234+0000 osd.3 (osd.3) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: cluster 2026-03-21T06:51:13.735247+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: cluster 2026-03-21T06:51:13.735247+0000 osd.3 (osd.3) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: audit 2026-03-21T06:51:15.431180+0000 mon.b (mon.2) 15 : audit [DBG] from='client.? 192.168.123.104:0/3495329696' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: audit 2026-03-21T06:51:15.431180+0000 mon.b (mon.2) 15 : audit [DBG] from='client.? 192.168.123.104:0/3495329696' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: cluster 2026-03-21T06:51:15.559550+0000 mgr.x (mgr.14152) 172 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: cluster 2026-03-21T06:51:15.559550+0000 mgr.x (mgr.14152) 172 : cluster [DBG] pgmap v110: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: cluster 2026-03-21T06:51:16.080201+0000 mon.a (mon.0) 492 : cluster [DBG] osdmap e28: 5 total, 4 up, 5 in 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: cluster 2026-03-21T06:51:16.080201+0000 mon.a (mon.0) 492 : cluster [DBG] osdmap e28: 5 total, 4 up, 5 in 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: audit 2026-03-21T06:51:16.080392+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:16.742 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:16 vm04 bash[20194]: audit 2026-03-21T06:51:16.080392+0000 mon.a (mon.0) 493 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:17.911 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:51:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:17.912 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:51:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:17.912 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: cluster 2026-03-21T06:51:17.083243+0000 mon.a (mon.0) 494 : cluster [DBG] osdmap e29: 5 total, 4 up, 5 in 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: cluster 2026-03-21T06:51:17.083243+0000 mon.a (mon.0) 494 : cluster [DBG] osdmap e29: 5 total, 4 up, 5 in 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: audit 2026-03-21T06:51:17.083381+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: audit 2026-03-21T06:51:17.083381+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: audit 2026-03-21T06:51:17.119556+0000 mon.a (mon.0) 496 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: audit 2026-03-21T06:51:17.119556+0000 mon.a (mon.0) 496 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: audit 2026-03-21T06:51:17.120215+0000 mon.a (mon.0) 497 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: audit 2026-03-21T06:51:17.120215+0000 mon.a (mon.0) 497 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: cephadm 2026-03-21T06:51:17.120703+0000 mgr.x (mgr.14152) 173 : cephadm [INF] Deploying daemon osd.4 on vm04 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: cephadm 2026-03-21T06:51:17.120703+0000 mgr.x (mgr.14152) 173 : cephadm [INF] Deploying daemon osd.4 on vm04 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: cluster 2026-03-21T06:51:17.559805+0000 mgr.x (mgr.14152) 174 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:18.240 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:18 vm04 bash[20194]: cluster 2026-03-21T06:51:17.559805+0000 mgr.x (mgr.14152) 174 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:18.240 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:51:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:18.240 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:51:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: cluster 2026-03-21T06:51:17.083243+0000 mon.a (mon.0) 494 : cluster [DBG] osdmap e29: 5 total, 4 up, 5 in 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: cluster 2026-03-21T06:51:17.083243+0000 mon.a (mon.0) 494 : cluster [DBG] osdmap e29: 5 total, 4 up, 5 in 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: audit 2026-03-21T06:51:17.083381+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: audit 2026-03-21T06:51:17.083381+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: audit 2026-03-21T06:51:17.119556+0000 mon.a (mon.0) 496 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: audit 2026-03-21T06:51:17.119556+0000 mon.a (mon.0) 496 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: audit 2026-03-21T06:51:17.120215+0000 mon.a (mon.0) 497 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: audit 2026-03-21T06:51:17.120215+0000 mon.a (mon.0) 497 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: cephadm 2026-03-21T06:51:17.120703+0000 mgr.x (mgr.14152) 173 : cephadm [INF] Deploying daemon osd.4 on vm04 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: cephadm 2026-03-21T06:51:17.120703+0000 mgr.x (mgr.14152) 173 : cephadm [INF] Deploying daemon osd.4 on vm04 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: cluster 2026-03-21T06:51:17.559805+0000 mgr.x (mgr.14152) 174 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:18.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:18 vm02 bash[17657]: cluster 2026-03-21T06:51:17.559805+0000 mgr.x (mgr.14152) 174 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: cluster 2026-03-21T06:51:17.083243+0000 mon.a (mon.0) 494 : cluster [DBG] osdmap e29: 5 total, 4 up, 5 in 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: cluster 2026-03-21T06:51:17.083243+0000 mon.a (mon.0) 494 : cluster [DBG] osdmap e29: 5 total, 4 up, 5 in 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: audit 2026-03-21T06:51:17.083381+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: audit 2026-03-21T06:51:17.083381+0000 mon.a (mon.0) 495 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: audit 2026-03-21T06:51:17.119556+0000 mon.a (mon.0) 496 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: audit 2026-03-21T06:51:17.119556+0000 mon.a (mon.0) 496 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.4"} : dispatch 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: audit 2026-03-21T06:51:17.120215+0000 mon.a (mon.0) 497 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: audit 2026-03-21T06:51:17.120215+0000 mon.a (mon.0) 497 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: cephadm 2026-03-21T06:51:17.120703+0000 mgr.x (mgr.14152) 173 : cephadm [INF] Deploying daemon osd.4 on vm04 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: cephadm 2026-03-21T06:51:17.120703+0000 mgr.x (mgr.14152) 173 : cephadm [INF] Deploying daemon osd.4 on vm04 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: cluster 2026-03-21T06:51:17.559805+0000 mgr.x (mgr.14152) 174 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:18.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:18 vm07 bash[19945]: cluster 2026-03-21T06:51:17.559805+0000 mgr.x (mgr.14152) 174 : cluster [DBG] pgmap v113: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:19.378 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:19 vm04 bash[20194]: audit 2026-03-21T06:51:18.114002+0000 mon.a (mon.0) 498 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.378 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:19 vm04 bash[20194]: audit 2026-03-21T06:51:18.114002+0000 mon.a (mon.0) 498 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.378 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:19 vm04 bash[20194]: audit 2026-03-21T06:51:18.118716+0000 mon.a (mon.0) 499 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.378 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:19 vm04 bash[20194]: audit 2026-03-21T06:51:18.118716+0000 mon.a (mon.0) 499 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:19 vm02 bash[17657]: audit 2026-03-21T06:51:18.114002+0000 mon.a (mon.0) 498 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:19 vm02 bash[17657]: audit 2026-03-21T06:51:18.114002+0000 mon.a (mon.0) 498 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:19 vm02 bash[17657]: audit 2026-03-21T06:51:18.118716+0000 mon.a (mon.0) 499 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.396 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:19 vm02 bash[17657]: audit 2026-03-21T06:51:18.118716+0000 mon.a (mon.0) 499 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:19 vm07 bash[19945]: audit 2026-03-21T06:51:18.114002+0000 mon.a (mon.0) 498 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:19 vm07 bash[19945]: audit 2026-03-21T06:51:18.114002+0000 mon.a (mon.0) 498 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:19 vm07 bash[19945]: audit 2026-03-21T06:51:18.118716+0000 mon.a (mon.0) 499 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:19 vm07 bash[19945]: audit 2026-03-21T06:51:18.118716+0000 mon.a (mon.0) 499 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:19.748 INFO:teuthology.orchestra.run.vm04.stdout:Created osd(s) 4 on host 'vm04' 2026-03-21T06:51:19.825 DEBUG:teuthology.orchestra.run.vm04:osd.4> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.4.service 2026-03-21T06:51:19.826 INFO:tasks.cephadm:Deploying osd.5 on vm07 with /dev/vg_nvme/lv_4... 2026-03-21T06:51:19.826 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- lvm zap /dev/vg_nvme/lv_4 2026-03-21T06:51:20.071 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:51:20.253 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:51:19 vm04 bash[33334]: debug 2026-03-21T06:51:19.955+0000 7f6930a578c0 -1 osd.4 0 log_to_monitors true 2026-03-21T06:51:20.566 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:51:20.582 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch daemon add osd vm07:vg_nvme/lv_4 --skip-validation 2026-03-21T06:51:20.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.518504+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.518504+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.524494+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.524494+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: cluster 2026-03-21T06:51:19.560079+0000 mgr.x (mgr.14152) 175 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:20.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: cluster 2026-03-21T06:51:19.560079+0000 mgr.x (mgr.14152) 175 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:20.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.736366+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.736366+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.741147+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.741147+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.798 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.956204+0000 mon.a (mon.0) 504 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:20.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.956204+0000 mon.a (mon.0) 504 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:20.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.964205+0000 mon.b (mon.2) 16 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:20.799 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:20 vm02 bash[17657]: audit 2026-03-21T06:51:19.964205+0000 mon.b (mon.2) 16 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:20.816 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:51:20.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.518504+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.518504+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.524494+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.524494+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: cluster 2026-03-21T06:51:19.560079+0000 mgr.x (mgr.14152) 175 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:20.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: cluster 2026-03-21T06:51:19.560079+0000 mgr.x (mgr.14152) 175 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:20.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.736366+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.736366+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.741147+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.741147+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:20.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.956204+0000 mon.a (mon.0) 504 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:20.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.956204+0000 mon.a (mon.0) 504 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:20.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.964205+0000 mon.b (mon.2) 16 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:20.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:20 vm07 bash[19945]: audit 2026-03-21T06:51:19.964205+0000 mon.b (mon.2) 16 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.518504+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.518504+0000 mon.a (mon.0) 500 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.524494+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.524494+0000 mon.a (mon.0) 501 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: cluster 2026-03-21T06:51:19.560079+0000 mgr.x (mgr.14152) 175 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: cluster 2026-03-21T06:51:19.560079+0000 mgr.x (mgr.14152) 175 : cluster [DBG] pgmap v114: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.736366+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.736366+0000 mon.a (mon.0) 502 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.741147+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.741147+0000 mon.a (mon.0) 503 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.956204+0000 mon.a (mon.0) 504 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.956204+0000 mon.a (mon.0) 504 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.964205+0000 mon.b (mon.2) 16 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:20 vm04 bash[20194]: audit 2026-03-21T06:51:19.964205+0000 mon.b (mon.2) 16 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]} : dispatch 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.589759+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.589759+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.595861+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.595861+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.882896+0000 mon.a (mon.0) 507 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.882896+0000 mon.a (mon.0) 507 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: cluster 2026-03-21T06:51:20.902058+0000 mon.a (mon.0) 508 : cluster [DBG] osdmap e30: 5 total, 4 up, 5 in 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: cluster 2026-03-21T06:51:20.902058+0000 mon.a (mon.0) 508 : cluster [DBG] osdmap e30: 5 total, 4 up, 5 in 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.902260+0000 mon.a (mon.0) 509 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.902260+0000 mon.a (mon.0) 509 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.904934+0000 mon.a (mon.0) 510 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.904934+0000 mon.a (mon.0) 510 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.913137+0000 mon.b (mon.2) 17 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:20.913137+0000 mon.b (mon.2) 17 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.197924+0000 mon.a (mon.0) 511 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.197924+0000 mon.a (mon.0) 511 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.201751+0000 mon.a (mon.0) 512 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.201751+0000 mon.a (mon.0) 512 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.202583+0000 mon.a (mon.0) 513 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:21.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.202583+0000 mon.a (mon.0) 513 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:21.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.204036+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:21.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.204036+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:21.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.204496+0000 mon.a (mon.0) 515 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:21.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.204496+0000 mon.a (mon.0) 515 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:21.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.513158+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.513158+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.517610+0000 mon.a (mon.0) 517 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:21 vm02 bash[17657]: audit 2026-03-21T06:51:21.517610+0000 mon.a (mon.0) 517 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.589759+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.589759+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.595861+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.595861+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.882896+0000 mon.a (mon.0) 507 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.882896+0000 mon.a (mon.0) 507 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: cluster 2026-03-21T06:51:20.902058+0000 mon.a (mon.0) 508 : cluster [DBG] osdmap e30: 5 total, 4 up, 5 in 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: cluster 2026-03-21T06:51:20.902058+0000 mon.a (mon.0) 508 : cluster [DBG] osdmap e30: 5 total, 4 up, 5 in 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.902260+0000 mon.a (mon.0) 509 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.902260+0000 mon.a (mon.0) 509 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.904934+0000 mon.a (mon.0) 510 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.904934+0000 mon.a (mon.0) 510 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.913137+0000 mon.b (mon.2) 17 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:20.913137+0000 mon.b (mon.2) 17 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.197924+0000 mon.a (mon.0) 511 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.197924+0000 mon.a (mon.0) 511 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.201751+0000 mon.a (mon.0) 512 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.201751+0000 mon.a (mon.0) 512 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.202583+0000 mon.a (mon.0) 513 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.202583+0000 mon.a (mon.0) 513 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.204036+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.204036+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.204496+0000 mon.a (mon.0) 515 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.204496+0000 mon.a (mon.0) 515 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.513158+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.513158+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.517610+0000 mon.a (mon.0) 517 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:21 vm07 bash[19945]: audit 2026-03-21T06:51:21.517610+0000 mon.a (mon.0) 517 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.589759+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.589759+0000 mon.a (mon.0) 505 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.595861+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.595861+0000 mon.a (mon.0) 506 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.882896+0000 mon.a (mon.0) 507 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.882896+0000 mon.a (mon.0) 507 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["4"]}]': finished 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: cluster 2026-03-21T06:51:20.902058+0000 mon.a (mon.0) 508 : cluster [DBG] osdmap e30: 5 total, 4 up, 5 in 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: cluster 2026-03-21T06:51:20.902058+0000 mon.a (mon.0) 508 : cluster [DBG] osdmap e30: 5 total, 4 up, 5 in 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.902260+0000 mon.a (mon.0) 509 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.902260+0000 mon.a (mon.0) 509 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.904934+0000 mon.a (mon.0) 510 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.904934+0000 mon.a (mon.0) 510 : audit [INF] from='osd.4 ' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.913137+0000 mon.b (mon.2) 17 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:20.913137+0000 mon.b (mon.2) 17 : audit [INF] from='osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955]' entity='osd.4' cmd={"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.197924+0000 mon.a (mon.0) 511 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.197924+0000 mon.a (mon.0) 511 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.201751+0000 mon.a (mon.0) 512 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.201751+0000 mon.a (mon.0) 512 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.202583+0000 mon.a (mon.0) 513 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.202583+0000 mon.a (mon.0) 513 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.204036+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.204036+0000 mon.a (mon.0) 514 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.204496+0000 mon.a (mon.0) 515 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.204496+0000 mon.a (mon.0) 515 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.513158+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.513158+0000 mon.a (mon.0) 516 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.517610+0000 mon.a (mon.0) 517 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:21.918 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:21 vm04 bash[20194]: audit 2026-03-21T06:51:21.517610+0000 mon.a (mon.0) 517 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.254 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:51:21 vm04 bash[33334]: debug 2026-03-21T06:51:21.919+0000 7f692c9d7640 -1 osd.4 0 waiting for initial osdmap 2026-03-21T06:51:22.254 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:51:21 vm04 bash[33334]: debug 2026-03-21T06:51:21.923+0000 7f6927fbf640 -1 osd.4 31 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:21.192260+0000 mgr.x (mgr.14152) 176 : audit [DBG] from='client.24241 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:21.192260+0000 mgr.x (mgr.14152) 176 : audit [DBG] from='client.24241 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cephadm 2026-03-21T06:51:21.193163+0000 mgr.x (mgr.14152) 177 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cephadm 2026-03-21T06:51:21.193163+0000 mgr.x (mgr.14152) 177 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cephadm 2026-03-21T06:51:21.193230+0000 mgr.x (mgr.14152) 178 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_4'] 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cephadm 2026-03-21T06:51:21.193230+0000 mgr.x (mgr.14152) 178 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_4'] 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cephadm 2026-03-21T06:51:21.198406+0000 mgr.x (mgr.14152) 179 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cephadm 2026-03-21T06:51:21.198406+0000 mgr.x (mgr.14152) 179 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cephadm 2026-03-21T06:51:21.198510+0000 mgr.x (mgr.14152) 180 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cephadm 2026-03-21T06:51:21.198510+0000 mgr.x (mgr.14152) 180 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cluster 2026-03-21T06:51:21.560319+0000 mgr.x (mgr.14152) 181 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cluster 2026-03-21T06:51:21.560319+0000 mgr.x (mgr.14152) 181 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:21.886111+0000 mon.a (mon.0) 518 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:21.886111+0000 mon.a (mon.0) 518 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cluster 2026-03-21T06:51:21.889042+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e31: 5 total, 4 up, 5 in 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cluster 2026-03-21T06:51:21.889042+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e31: 5 total, 4 up, 5 in 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:21.889242+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:21.889242+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:21.893089+0000 mon.a (mon.0) 521 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:21.893089+0000 mon.a (mon.0) 521 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.057067+0000 mon.a (mon.0) 522 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.057067+0000 mon.a (mon.0) 522 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.060658+0000 mon.a (mon.0) 523 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"}]': finished 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.060658+0000 mon.a (mon.0) 523 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"}]': finished 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.061910+0000 mon.c (mon.1) 8 : audit [INF] from='client.? 192.168.123.107:0/3843394450' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.061910+0000 mon.c (mon.1) 8 : audit [INF] from='client.? 192.168.123.107:0/3843394450' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cluster 2026-03-21T06:51:22.063661+0000 mon.a (mon.0) 524 : cluster [INF] osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955] boot 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cluster 2026-03-21T06:51:22.063661+0000 mon.a (mon.0) 524 : cluster [INF] osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955] boot 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cluster 2026-03-21T06:51:22.063693+0000 mon.a (mon.0) 525 : cluster [DBG] osdmap e32: 6 total, 5 up, 6 in 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: cluster 2026-03-21T06:51:22.063693+0000 mon.a (mon.0) 525 : cluster [DBG] osdmap e32: 6 total, 5 up, 6 in 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.064640+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.064640+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.064927+0000 mon.a (mon.0) 527 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.064927+0000 mon.a (mon.0) 527 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.431088+0000 mon.c (mon.1) 9 : audit [DBG] from='client.? 192.168.123.107:0/3734853751' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.431088+0000 mon.c (mon.1) 9 : audit [DBG] from='client.? 192.168.123.107:0/3734853751' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.543540+0000 mon.a (mon.0) 528 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.754 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.543540+0000 mon.a (mon.0) 528 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.548514+0000 mon.a (mon.0) 529 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.548514+0000 mon.a (mon.0) 529 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.553417+0000 mon.a (mon.0) 530 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.553417+0000 mon.a (mon.0) 530 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.558259+0000 mon.a (mon.0) 531 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.558259+0000 mon.a (mon.0) 531 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.558898+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.558898+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.559402+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.559402+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.562466+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.562466+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.563927+0000 mon.a (mon.0) 535 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.563927+0000 mon.a (mon.0) 535 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.564436+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.564436+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.568771+0000 mon.a (mon.0) 537 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.568771+0000 mon.a (mon.0) 537 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.569970+0000 mon.a (mon.0) 538 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.569970+0000 mon.a (mon.0) 538 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.571216+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.571216+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.571647+0000 mon.a (mon.0) 540 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:22.755 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:22 vm04 bash[20194]: audit 2026-03-21T06:51:22.571647+0000 mon.a (mon.0) 540 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:21.192260+0000 mgr.x (mgr.14152) 176 : audit [DBG] from='client.24241 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:21.192260+0000 mgr.x (mgr.14152) 176 : audit [DBG] from='client.24241 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cephadm 2026-03-21T06:51:21.193163+0000 mgr.x (mgr.14152) 177 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cephadm 2026-03-21T06:51:21.193163+0000 mgr.x (mgr.14152) 177 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cephadm 2026-03-21T06:51:21.193230+0000 mgr.x (mgr.14152) 178 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_4'] 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cephadm 2026-03-21T06:51:21.193230+0000 mgr.x (mgr.14152) 178 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_4'] 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cephadm 2026-03-21T06:51:21.198406+0000 mgr.x (mgr.14152) 179 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cephadm 2026-03-21T06:51:21.198406+0000 mgr.x (mgr.14152) 179 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cephadm 2026-03-21T06:51:21.198510+0000 mgr.x (mgr.14152) 180 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cephadm 2026-03-21T06:51:21.198510+0000 mgr.x (mgr.14152) 180 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cluster 2026-03-21T06:51:21.560319+0000 mgr.x (mgr.14152) 181 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cluster 2026-03-21T06:51:21.560319+0000 mgr.x (mgr.14152) 181 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:21.886111+0000 mon.a (mon.0) 518 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:21.886111+0000 mon.a (mon.0) 518 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:22.881 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cluster 2026-03-21T06:51:21.889042+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e31: 5 total, 4 up, 5 in 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cluster 2026-03-21T06:51:21.889042+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e31: 5 total, 4 up, 5 in 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:21.889242+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:21.889242+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:21.893089+0000 mon.a (mon.0) 521 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:21.893089+0000 mon.a (mon.0) 521 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.057067+0000 mon.a (mon.0) 522 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.057067+0000 mon.a (mon.0) 522 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.060658+0000 mon.a (mon.0) 523 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"}]': finished 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.060658+0000 mon.a (mon.0) 523 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"}]': finished 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.061910+0000 mon.c (mon.1) 8 : audit [INF] from='client.? 192.168.123.107:0/3843394450' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.061910+0000 mon.c (mon.1) 8 : audit [INF] from='client.? 192.168.123.107:0/3843394450' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cluster 2026-03-21T06:51:22.063661+0000 mon.a (mon.0) 524 : cluster [INF] osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955] boot 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cluster 2026-03-21T06:51:22.063661+0000 mon.a (mon.0) 524 : cluster [INF] osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955] boot 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cluster 2026-03-21T06:51:22.063693+0000 mon.a (mon.0) 525 : cluster [DBG] osdmap e32: 6 total, 5 up, 6 in 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: cluster 2026-03-21T06:51:22.063693+0000 mon.a (mon.0) 525 : cluster [DBG] osdmap e32: 6 total, 5 up, 6 in 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.064640+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.064640+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.064927+0000 mon.a (mon.0) 527 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.064927+0000 mon.a (mon.0) 527 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.431088+0000 mon.c (mon.1) 9 : audit [DBG] from='client.? 192.168.123.107:0/3734853751' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.431088+0000 mon.c (mon.1) 9 : audit [DBG] from='client.? 192.168.123.107:0/3734853751' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.543540+0000 mon.a (mon.0) 528 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.543540+0000 mon.a (mon.0) 528 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.548514+0000 mon.a (mon.0) 529 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.548514+0000 mon.a (mon.0) 529 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.553417+0000 mon.a (mon.0) 530 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.553417+0000 mon.a (mon.0) 530 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.558259+0000 mon.a (mon.0) 531 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.558259+0000 mon.a (mon.0) 531 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.558898+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.558898+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.559402+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.559402+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.562466+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.562466+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.563927+0000 mon.a (mon.0) 535 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.563927+0000 mon.a (mon.0) 535 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.564436+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.564436+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.568771+0000 mon.a (mon.0) 537 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.568771+0000 mon.a (mon.0) 537 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.569970+0000 mon.a (mon.0) 538 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.569970+0000 mon.a (mon.0) 538 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.571216+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.571216+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.571647+0000 mon.a (mon.0) 540 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:22.882 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:22 vm07 bash[19945]: audit 2026-03-21T06:51:22.571647+0000 mon.a (mon.0) 540 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:21.192260+0000 mgr.x (mgr.14152) 176 : audit [DBG] from='client.24241 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:21.192260+0000 mgr.x (mgr.14152) 176 : audit [DBG] from='client.24241 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_4", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cephadm 2026-03-21T06:51:21.193163+0000 mgr.x (mgr.14152) 177 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cephadm 2026-03-21T06:51:21.193163+0000 mgr.x (mgr.14152) 177 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cephadm 2026-03-21T06:51:21.193230+0000 mgr.x (mgr.14152) 178 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_4'] 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cephadm 2026-03-21T06:51:21.193230+0000 mgr.x (mgr.14152) 178 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_4'] 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cephadm 2026-03-21T06:51:21.198406+0000 mgr.x (mgr.14152) 179 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cephadm 2026-03-21T06:51:21.198406+0000 mgr.x (mgr.14152) 179 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cephadm 2026-03-21T06:51:21.198510+0000 mgr.x (mgr.14152) 180 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cephadm 2026-03-21T06:51:21.198510+0000 mgr.x (mgr.14152) 180 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cluster 2026-03-21T06:51:21.560319+0000 mgr.x (mgr.14152) 181 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cluster 2026-03-21T06:51:21.560319+0000 mgr.x (mgr.14152) 181 : cluster [DBG] pgmap v116: 1 pgs: 1 active+clean; 449 KiB data, 108 MiB used, 80 GiB / 80 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:21.886111+0000 mon.a (mon.0) 518 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:21.886111+0000 mon.a (mon.0) 518 : audit [INF] from='osd.4 ' entity='osd.4' cmd='[{"prefix": "osd crush create-or-move", "id": 4, "weight":0.0195, "args": ["host=vm04", "root=default"]}]': finished 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cluster 2026-03-21T06:51:21.889042+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e31: 5 total, 4 up, 5 in 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cluster 2026-03-21T06:51:21.889042+0000 mon.a (mon.0) 519 : cluster [DBG] osdmap e31: 5 total, 4 up, 5 in 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:21.889242+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:21.889242+0000 mon.a (mon.0) 520 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:21.893089+0000 mon.a (mon.0) 521 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:21.893089+0000 mon.a (mon.0) 521 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.057067+0000 mon.a (mon.0) 522 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.057067+0000 mon.a (mon.0) 522 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.060658+0000 mon.a (mon.0) 523 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"}]': finished 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.060658+0000 mon.a (mon.0) 523 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"}]': finished 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.061910+0000 mon.c (mon.1) 8 : audit [INF] from='client.? 192.168.123.107:0/3843394450' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.061910+0000 mon.c (mon.1) 8 : audit [INF] from='client.? 192.168.123.107:0/3843394450' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "a697bb58-dae5-4cba-b085-8e5769734d61"} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cluster 2026-03-21T06:51:22.063661+0000 mon.a (mon.0) 524 : cluster [INF] osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955] boot 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cluster 2026-03-21T06:51:22.063661+0000 mon.a (mon.0) 524 : cluster [INF] osd.4 [v2:192.168.123.104:6816/3736423955,v1:192.168.123.104:6817/3736423955] boot 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cluster 2026-03-21T06:51:22.063693+0000 mon.a (mon.0) 525 : cluster [DBG] osdmap e32: 6 total, 5 up, 6 in 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: cluster 2026-03-21T06:51:22.063693+0000 mon.a (mon.0) 525 : cluster [DBG] osdmap e32: 6 total, 5 up, 6 in 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.064640+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.064640+0000 mon.a (mon.0) 526 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 4} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.064927+0000 mon.a (mon.0) 527 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.064927+0000 mon.a (mon.0) 527 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.431088+0000 mon.c (mon.1) 9 : audit [DBG] from='client.? 192.168.123.107:0/3734853751' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.431088+0000 mon.c (mon.1) 9 : audit [DBG] from='client.? 192.168.123.107:0/3734853751' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.543540+0000 mon.a (mon.0) 528 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.543540+0000 mon.a (mon.0) 528 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.548514+0000 mon.a (mon.0) 529 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.548514+0000 mon.a (mon.0) 529 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.553417+0000 mon.a (mon.0) 530 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.897 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.553417+0000 mon.a (mon.0) 530 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.558259+0000 mon.a (mon.0) 531 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.558259+0000 mon.a (mon.0) 531 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.558898+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.558898+0000 mon.a (mon.0) 532 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.559402+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.559402+0000 mon.a (mon.0) 533 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.562466+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.562466+0000 mon.a (mon.0) 534 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.563927+0000 mon.a (mon.0) 535 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.563927+0000 mon.a (mon.0) 535 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.564436+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.564436+0000 mon.a (mon.0) 536 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.568771+0000 mon.a (mon.0) 537 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.568771+0000 mon.a (mon.0) 537 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.569970+0000 mon.a (mon.0) 538 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.569970+0000 mon.a (mon.0) 538 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.571216+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.571216+0000 mon.a (mon.0) 539 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.571647+0000 mon.a (mon.0) 540 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:22.898 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:22 vm02 bash[17657]: audit 2026-03-21T06:51:22.571647+0000 mon.a (mon.0) 540 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:23.810 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:23 vm07 bash[19945]: cluster 2026-03-21T06:51:20.945244+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:23.810 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:23 vm07 bash[19945]: cluster 2026-03-21T06:51:20.945244+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:23.810 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:23 vm07 bash[19945]: cluster 2026-03-21T06:51:20.945257+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:23.810 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:23 vm07 bash[19945]: cluster 2026-03-21T06:51:20.945257+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:23.810 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:23 vm07 bash[19945]: cephadm 2026-03-21T06:51:22.537223+0000 mgr.x (mgr.14152) 182 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:23.810 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:23 vm07 bash[19945]: cephadm 2026-03-21T06:51:22.537223+0000 mgr.x (mgr.14152) 182 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:23.810 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:23 vm07 bash[19945]: cephadm 2026-03-21T06:51:22.559794+0000 mgr.x (mgr.14152) 183 : cephadm [INF] Adjusting osd_memory_target on vm04 to 2967M 2026-03-21T06:51:23.810 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:23 vm07 bash[19945]: cephadm 2026-03-21T06:51:22.559794+0000 mgr.x (mgr.14152) 183 : cephadm [INF] Adjusting osd_memory_target on vm04 to 2967M 2026-03-21T06:51:23.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:23 vm02 bash[17657]: cluster 2026-03-21T06:51:20.945244+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:23.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:23 vm02 bash[17657]: cluster 2026-03-21T06:51:20.945244+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:23.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:23 vm02 bash[17657]: cluster 2026-03-21T06:51:20.945257+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:23.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:23 vm02 bash[17657]: cluster 2026-03-21T06:51:20.945257+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:23.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:23 vm02 bash[17657]: cephadm 2026-03-21T06:51:22.537223+0000 mgr.x (mgr.14152) 182 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:23.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:23 vm02 bash[17657]: cephadm 2026-03-21T06:51:22.537223+0000 mgr.x (mgr.14152) 182 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:23.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:23 vm02 bash[17657]: cephadm 2026-03-21T06:51:22.559794+0000 mgr.x (mgr.14152) 183 : cephadm [INF] Adjusting osd_memory_target on vm04 to 2967M 2026-03-21T06:51:23.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:23 vm02 bash[17657]: cephadm 2026-03-21T06:51:22.559794+0000 mgr.x (mgr.14152) 183 : cephadm [INF] Adjusting osd_memory_target on vm04 to 2967M 2026-03-21T06:51:24.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:23 vm04 bash[20194]: cluster 2026-03-21T06:51:20.945244+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:24.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:23 vm04 bash[20194]: cluster 2026-03-21T06:51:20.945244+0000 osd.4 (osd.4) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:24.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:23 vm04 bash[20194]: cluster 2026-03-21T06:51:20.945257+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:24.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:23 vm04 bash[20194]: cluster 2026-03-21T06:51:20.945257+0000 osd.4 (osd.4) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:24.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:23 vm04 bash[20194]: cephadm 2026-03-21T06:51:22.537223+0000 mgr.x (mgr.14152) 182 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:24.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:23 vm04 bash[20194]: cephadm 2026-03-21T06:51:22.537223+0000 mgr.x (mgr.14152) 182 : cephadm [INF] Detected new or changed devices on vm04 2026-03-21T06:51:24.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:23 vm04 bash[20194]: cephadm 2026-03-21T06:51:22.559794+0000 mgr.x (mgr.14152) 183 : cephadm [INF] Adjusting osd_memory_target on vm04 to 2967M 2026-03-21T06:51:24.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:23 vm04 bash[20194]: cephadm 2026-03-21T06:51:22.559794+0000 mgr.x (mgr.14152) 183 : cephadm [INF] Adjusting osd_memory_target on vm04 to 2967M 2026-03-21T06:51:24.835 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: cluster 2026-03-21T06:51:23.560655+0000 mgr.x (mgr.14152) 184 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:24.835 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: cluster 2026-03-21T06:51:23.560655+0000 mgr.x (mgr.14152) 184 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:24.835 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: cluster 2026-03-21T06:51:23.603227+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e33: 6 total, 5 up, 6 in 2026-03-21T06:51:24.836 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: cluster 2026-03-21T06:51:23.603227+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e33: 6 total, 5 up, 6 in 2026-03-21T06:51:24.836 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: audit 2026-03-21T06:51:23.603322+0000 mon.a (mon.0) 542 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:24.836 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: audit 2026-03-21T06:51:23.603322+0000 mon.a (mon.0) 542 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:24.836 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: audit 2026-03-21T06:51:24.264635+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:24.836 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: audit 2026-03-21T06:51:24.264635+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:24.836 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: audit 2026-03-21T06:51:24.265287+0000 mon.a (mon.0) 544 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:24.836 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: audit 2026-03-21T06:51:24.265287+0000 mon.a (mon.0) 544 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:24.836 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: audit 2026-03-21T06:51:24.585152+0000 mon.a (mon.0) 545 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:24.836 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: audit 2026-03-21T06:51:24.585152+0000 mon.a (mon.0) 545 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:24.836 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: audit 2026-03-21T06:51:24.585753+0000 mon.a (mon.0) 546 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:24.836 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:24 vm07 bash[19945]: audit 2026-03-21T06:51:24.585753+0000 mon.a (mon.0) 546 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: cluster 2026-03-21T06:51:23.560655+0000 mgr.x (mgr.14152) 184 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: cluster 2026-03-21T06:51:23.560655+0000 mgr.x (mgr.14152) 184 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: cluster 2026-03-21T06:51:23.603227+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e33: 6 total, 5 up, 6 in 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: cluster 2026-03-21T06:51:23.603227+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e33: 6 total, 5 up, 6 in 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: audit 2026-03-21T06:51:23.603322+0000 mon.a (mon.0) 542 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: audit 2026-03-21T06:51:23.603322+0000 mon.a (mon.0) 542 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: audit 2026-03-21T06:51:24.264635+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: audit 2026-03-21T06:51:24.264635+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: audit 2026-03-21T06:51:24.265287+0000 mon.a (mon.0) 544 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: audit 2026-03-21T06:51:24.265287+0000 mon.a (mon.0) 544 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: audit 2026-03-21T06:51:24.585152+0000 mon.a (mon.0) 545 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: audit 2026-03-21T06:51:24.585152+0000 mon.a (mon.0) 545 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: audit 2026-03-21T06:51:24.585753+0000 mon.a (mon.0) 546 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:24.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:24 vm02 bash[17657]: audit 2026-03-21T06:51:24.585753+0000 mon.a (mon.0) 546 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:25.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: cluster 2026-03-21T06:51:23.560655+0000 mgr.x (mgr.14152) 184 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:25.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: cluster 2026-03-21T06:51:23.560655+0000 mgr.x (mgr.14152) 184 : cluster [DBG] pgmap v119: 1 pgs: 1 active+clean; 449 KiB data, 134 MiB used, 100 GiB / 100 GiB avail; 75 KiB/s, 0 objects/s recovering 2026-03-21T06:51:25.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: cluster 2026-03-21T06:51:23.603227+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e33: 6 total, 5 up, 6 in 2026-03-21T06:51:25.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: cluster 2026-03-21T06:51:23.603227+0000 mon.a (mon.0) 541 : cluster [DBG] osdmap e33: 6 total, 5 up, 6 in 2026-03-21T06:51:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: audit 2026-03-21T06:51:23.603322+0000 mon.a (mon.0) 542 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: audit 2026-03-21T06:51:23.603322+0000 mon.a (mon.0) 542 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: audit 2026-03-21T06:51:24.264635+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: audit 2026-03-21T06:51:24.264635+0000 mon.a (mon.0) 543 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: audit 2026-03-21T06:51:24.265287+0000 mon.a (mon.0) 544 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: audit 2026-03-21T06:51:24.265287+0000 mon.a (mon.0) 544 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: audit 2026-03-21T06:51:24.585152+0000 mon.a (mon.0) 545 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: audit 2026-03-21T06:51:24.585152+0000 mon.a (mon.0) 545 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.5"} : dispatch 2026-03-21T06:51:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: audit 2026-03-21T06:51:24.585753+0000 mon.a (mon.0) 546 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:24 vm04 bash[20194]: audit 2026-03-21T06:51:24.585753+0000 mon.a (mon.0) 546 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:51:25.350 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:25 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:25.350 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:25 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:25.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:25 vm02 bash[17657]: cephadm 2026-03-21T06:51:24.265716+0000 mgr.x (mgr.14152) 185 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:25.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:25 vm02 bash[17657]: cephadm 2026-03-21T06:51:24.265716+0000 mgr.x (mgr.14152) 185 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:25.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:25 vm02 bash[17657]: cephadm 2026-03-21T06:51:24.586209+0000 mgr.x (mgr.14152) 186 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:25.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:25 vm02 bash[17657]: cephadm 2026-03-21T06:51:24.586209+0000 mgr.x (mgr.14152) 186 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:25.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:25 vm02 bash[17657]: audit 2026-03-21T06:51:25.373801+0000 mon.a (mon.0) 547 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:25.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:25 vm02 bash[17657]: audit 2026-03-21T06:51:25.373801+0000 mon.a (mon.0) 547 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:25.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:25 vm02 bash[17657]: audit 2026-03-21T06:51:25.379983+0000 mon.a (mon.0) 548 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:25.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:25 vm02 bash[17657]: audit 2026-03-21T06:51:25.379983+0000 mon.a (mon.0) 548 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:25.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:25 vm07 bash[19945]: cephadm 2026-03-21T06:51:24.265716+0000 mgr.x (mgr.14152) 185 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:25.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:25 vm07 bash[19945]: cephadm 2026-03-21T06:51:24.265716+0000 mgr.x (mgr.14152) 185 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:25.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:25 vm07 bash[19945]: cephadm 2026-03-21T06:51:24.586209+0000 mgr.x (mgr.14152) 186 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:25.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:25 vm07 bash[19945]: cephadm 2026-03-21T06:51:24.586209+0000 mgr.x (mgr.14152) 186 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:25.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:25 vm07 bash[19945]: audit 2026-03-21T06:51:25.373801+0000 mon.a (mon.0) 547 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:25.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:25 vm07 bash[19945]: audit 2026-03-21T06:51:25.373801+0000 mon.a (mon.0) 547 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:25.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:25 vm07 bash[19945]: audit 2026-03-21T06:51:25.379983+0000 mon.a (mon.0) 548 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:25.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:25 vm07 bash[19945]: audit 2026-03-21T06:51:25.379983+0000 mon.a (mon.0) 548 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:26.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:25 vm04 bash[20194]: cephadm 2026-03-21T06:51:24.265716+0000 mgr.x (mgr.14152) 185 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:26.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:25 vm04 bash[20194]: cephadm 2026-03-21T06:51:24.265716+0000 mgr.x (mgr.14152) 185 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:26.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:25 vm04 bash[20194]: cephadm 2026-03-21T06:51:24.586209+0000 mgr.x (mgr.14152) 186 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:26.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:25 vm04 bash[20194]: cephadm 2026-03-21T06:51:24.586209+0000 mgr.x (mgr.14152) 186 : cephadm [INF] Deploying daemon osd.5 on vm07 2026-03-21T06:51:26.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:25 vm04 bash[20194]: audit 2026-03-21T06:51:25.373801+0000 mon.a (mon.0) 547 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:26.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:25 vm04 bash[20194]: audit 2026-03-21T06:51:25.373801+0000 mon.a (mon.0) 547 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:26.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:25 vm04 bash[20194]: audit 2026-03-21T06:51:25.379983+0000 mon.a (mon.0) 548 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:26.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:25 vm04 bash[20194]: audit 2026-03-21T06:51:25.379983+0000 mon.a (mon.0) 548 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:51:26.197 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:26 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:51:26.885 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:26 vm07 bash[19945]: cluster 2026-03-21T06:51:25.560937+0000 mgr.x (mgr.14152) 187 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:26.885 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:26 vm07 bash[19945]: cluster 2026-03-21T06:51:25.560937+0000 mgr.x (mgr.14152) 187 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:26.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:26 vm02 bash[17657]: cluster 2026-03-21T06:51:25.560937+0000 mgr.x (mgr.14152) 187 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:26.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:26 vm02 bash[17657]: cluster 2026-03-21T06:51:25.560937+0000 mgr.x (mgr.14152) 187 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:27.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:26 vm04 bash[20194]: cluster 2026-03-21T06:51:25.560937+0000 mgr.x (mgr.14152) 187 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:27.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:26 vm04 bash[20194]: cluster 2026-03-21T06:51:25.560937+0000 mgr.x (mgr.14152) 187 : cluster [DBG] pgmap v121: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:28.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:28 vm02 bash[17657]: cluster 2026-03-21T06:51:27.561219+0000 mgr.x (mgr.14152) 188 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:28.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:28 vm02 bash[17657]: cluster 2026-03-21T06:51:27.561219+0000 mgr.x (mgr.14152) 188 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:28.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:28 vm02 bash[17657]: audit 2026-03-21T06:51:27.619216+0000 mon.a (mon.0) 549 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:28.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:28 vm02 bash[17657]: audit 2026-03-21T06:51:27.619216+0000 mon.a (mon.0) 549 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:28.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:28 vm02 bash[17657]: audit 2026-03-21T06:51:27.624212+0000 mon.c (mon.1) 10 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:28.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:28 vm02 bash[17657]: audit 2026-03-21T06:51:27.624212+0000 mon.c (mon.1) 10 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:28.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:28 vm07 bash[19945]: cluster 2026-03-21T06:51:27.561219+0000 mgr.x (mgr.14152) 188 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:28.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:28 vm07 bash[19945]: cluster 2026-03-21T06:51:27.561219+0000 mgr.x (mgr.14152) 188 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:28.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:28 vm07 bash[19945]: audit 2026-03-21T06:51:27.619216+0000 mon.a (mon.0) 549 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:28.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:28 vm07 bash[19945]: audit 2026-03-21T06:51:27.619216+0000 mon.a (mon.0) 549 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:28.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:28 vm07 bash[19945]: audit 2026-03-21T06:51:27.624212+0000 mon.c (mon.1) 10 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:28.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:28 vm07 bash[19945]: audit 2026-03-21T06:51:27.624212+0000 mon.c (mon.1) 10 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:29.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:28 vm04 bash[20194]: cluster 2026-03-21T06:51:27.561219+0000 mgr.x (mgr.14152) 188 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:29.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:28 vm04 bash[20194]: cluster 2026-03-21T06:51:27.561219+0000 mgr.x (mgr.14152) 188 : cluster [DBG] pgmap v122: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:29.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:28 vm04 bash[20194]: audit 2026-03-21T06:51:27.619216+0000 mon.a (mon.0) 549 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:29.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:28 vm04 bash[20194]: audit 2026-03-21T06:51:27.619216+0000 mon.a (mon.0) 549 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:29.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:28 vm04 bash[20194]: audit 2026-03-21T06:51:27.624212+0000 mon.c (mon.1) 10 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:29.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:28 vm04 bash[20194]: audit 2026-03-21T06:51:27.624212+0000 mon.c (mon.1) 10 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:51:29.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:29 vm02 bash[17657]: audit 2026-03-21T06:51:28.619364+0000 mon.a (mon.0) 550 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:51:29.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:29 vm02 bash[17657]: audit 2026-03-21T06:51:28.619364+0000 mon.a (mon.0) 550 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:51:29.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:29 vm02 bash[17657]: cluster 2026-03-21T06:51:28.621080+0000 mon.a (mon.0) 551 : cluster [DBG] osdmap e34: 6 total, 5 up, 6 in 2026-03-21T06:51:29.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:29 vm02 bash[17657]: cluster 2026-03-21T06:51:28.621080+0000 mon.a (mon.0) 551 : cluster [DBG] osdmap e34: 6 total, 5 up, 6 in 2026-03-21T06:51:29.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:29 vm02 bash[17657]: audit 2026-03-21T06:51:28.621294+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:29.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:29 vm02 bash[17657]: audit 2026-03-21T06:51:28.621294+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:29.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:29 vm02 bash[17657]: audit 2026-03-21T06:51:28.622570+0000 mon.a (mon.0) 553 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:29.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:29 vm02 bash[17657]: audit 2026-03-21T06:51:28.622570+0000 mon.a (mon.0) 553 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:29.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:29 vm02 bash[17657]: audit 2026-03-21T06:51:28.627506+0000 mon.c (mon.1) 11 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:29.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:29 vm02 bash[17657]: audit 2026-03-21T06:51:28.627506+0000 mon.c (mon.1) 11 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:29.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:29 vm07 bash[19945]: audit 2026-03-21T06:51:28.619364+0000 mon.a (mon.0) 550 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:51:29.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:29 vm07 bash[19945]: audit 2026-03-21T06:51:28.619364+0000 mon.a (mon.0) 550 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:51:29.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:29 vm07 bash[19945]: cluster 2026-03-21T06:51:28.621080+0000 mon.a (mon.0) 551 : cluster [DBG] osdmap e34: 6 total, 5 up, 6 in 2026-03-21T06:51:29.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:29 vm07 bash[19945]: cluster 2026-03-21T06:51:28.621080+0000 mon.a (mon.0) 551 : cluster [DBG] osdmap e34: 6 total, 5 up, 6 in 2026-03-21T06:51:29.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:29 vm07 bash[19945]: audit 2026-03-21T06:51:28.621294+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:29.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:29 vm07 bash[19945]: audit 2026-03-21T06:51:28.621294+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:29.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:29 vm07 bash[19945]: audit 2026-03-21T06:51:28.622570+0000 mon.a (mon.0) 553 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:29.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:29 vm07 bash[19945]: audit 2026-03-21T06:51:28.622570+0000 mon.a (mon.0) 553 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:29.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:29 vm07 bash[19945]: audit 2026-03-21T06:51:28.627506+0000 mon.c (mon.1) 11 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:29.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:29 vm07 bash[19945]: audit 2026-03-21T06:51:28.627506+0000 mon.c (mon.1) 11 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:30.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:29 vm04 bash[20194]: audit 2026-03-21T06:51:28.619364+0000 mon.a (mon.0) 550 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:51:30.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:29 vm04 bash[20194]: audit 2026-03-21T06:51:28.619364+0000 mon.a (mon.0) 550 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:51:30.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:29 vm04 bash[20194]: cluster 2026-03-21T06:51:28.621080+0000 mon.a (mon.0) 551 : cluster [DBG] osdmap e34: 6 total, 5 up, 6 in 2026-03-21T06:51:30.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:29 vm04 bash[20194]: cluster 2026-03-21T06:51:28.621080+0000 mon.a (mon.0) 551 : cluster [DBG] osdmap e34: 6 total, 5 up, 6 in 2026-03-21T06:51:30.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:29 vm04 bash[20194]: audit 2026-03-21T06:51:28.621294+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:29 vm04 bash[20194]: audit 2026-03-21T06:51:28.621294+0000 mon.a (mon.0) 552 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:29 vm04 bash[20194]: audit 2026-03-21T06:51:28.622570+0000 mon.a (mon.0) 553 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:30.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:29 vm04 bash[20194]: audit 2026-03-21T06:51:28.622570+0000 mon.a (mon.0) 553 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:30.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:29 vm04 bash[20194]: audit 2026-03-21T06:51:28.627506+0000 mon.c (mon.1) 11 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:30.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:29 vm04 bash[20194]: audit 2026-03-21T06:51:28.627506+0000 mon.c (mon.1) 11 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: cluster 2026-03-21T06:51:28.648527+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: cluster 2026-03-21T06:51:28.648527+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: cluster 2026-03-21T06:51:28.648537+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: cluster 2026-03-21T06:51:28.648537+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: cluster 2026-03-21T06:51:29.561607+0000 mgr.x (mgr.14152) 189 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: cluster 2026-03-21T06:51:29.561607+0000 mgr.x (mgr.14152) 189 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: audit 2026-03-21T06:51:29.623227+0000 mon.a (mon.0) 554 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: audit 2026-03-21T06:51:29.623227+0000 mon.a (mon.0) 554 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: cluster 2026-03-21T06:51:29.625539+0000 mon.a (mon.0) 555 : cluster [DBG] osdmap e35: 6 total, 5 up, 6 in 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: cluster 2026-03-21T06:51:29.625539+0000 mon.a (mon.0) 555 : cluster [DBG] osdmap e35: 6 total, 5 up, 6 in 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: audit 2026-03-21T06:51:29.626431+0000 mon.a (mon.0) 556 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: audit 2026-03-21T06:51:29.626431+0000 mon.a (mon.0) 556 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: audit 2026-03-21T06:51:29.629162+0000 mon.a (mon.0) 557 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: audit 2026-03-21T06:51:29.629162+0000 mon.a (mon.0) 557 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: audit 2026-03-21T06:51:30.629067+0000 mon.a (mon.0) 558 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:30 vm02 bash[17657]: audit 2026-03-21T06:51:30.629067+0000 mon.a (mon.0) 558 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: cluster 2026-03-21T06:51:28.648527+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: cluster 2026-03-21T06:51:28.648527+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: cluster 2026-03-21T06:51:28.648537+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: cluster 2026-03-21T06:51:28.648537+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: cluster 2026-03-21T06:51:29.561607+0000 mgr.x (mgr.14152) 189 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: cluster 2026-03-21T06:51:29.561607+0000 mgr.x (mgr.14152) 189 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: audit 2026-03-21T06:51:29.623227+0000 mon.a (mon.0) 554 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: audit 2026-03-21T06:51:29.623227+0000 mon.a (mon.0) 554 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: cluster 2026-03-21T06:51:29.625539+0000 mon.a (mon.0) 555 : cluster [DBG] osdmap e35: 6 total, 5 up, 6 in 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: cluster 2026-03-21T06:51:29.625539+0000 mon.a (mon.0) 555 : cluster [DBG] osdmap e35: 6 total, 5 up, 6 in 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: audit 2026-03-21T06:51:29.626431+0000 mon.a (mon.0) 556 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: audit 2026-03-21T06:51:29.626431+0000 mon.a (mon.0) 556 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: audit 2026-03-21T06:51:29.629162+0000 mon.a (mon.0) 557 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: audit 2026-03-21T06:51:29.629162+0000 mon.a (mon.0) 557 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: audit 2026-03-21T06:51:30.629067+0000 mon.a (mon.0) 558 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:30.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:30 vm07 bash[19945]: audit 2026-03-21T06:51:30.629067+0000 mon.a (mon.0) 558 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:31.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: cluster 2026-03-21T06:51:28.648527+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: cluster 2026-03-21T06:51:28.648527+0000 osd.5 (osd.5) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: cluster 2026-03-21T06:51:28.648537+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: cluster 2026-03-21T06:51:28.648537+0000 osd.5 (osd.5) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: cluster 2026-03-21T06:51:29.561607+0000 mgr.x (mgr.14152) 189 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: cluster 2026-03-21T06:51:29.561607+0000 mgr.x (mgr.14152) 189 : cluster [DBG] pgmap v124: 1 pgs: 1 active+clean; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: audit 2026-03-21T06:51:29.623227+0000 mon.a (mon.0) 554 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: audit 2026-03-21T06:51:29.623227+0000 mon.a (mon.0) 554 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: cluster 2026-03-21T06:51:29.625539+0000 mon.a (mon.0) 555 : cluster [DBG] osdmap e35: 6 total, 5 up, 6 in 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: cluster 2026-03-21T06:51:29.625539+0000 mon.a (mon.0) 555 : cluster [DBG] osdmap e35: 6 total, 5 up, 6 in 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: audit 2026-03-21T06:51:29.626431+0000 mon.a (mon.0) 556 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: audit 2026-03-21T06:51:29.626431+0000 mon.a (mon.0) 556 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: audit 2026-03-21T06:51:29.629162+0000 mon.a (mon.0) 557 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: audit 2026-03-21T06:51:29.629162+0000 mon.a (mon.0) 557 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: audit 2026-03-21T06:51:30.629067+0000 mon.a (mon.0) 558 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:31.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:30 vm04 bash[20194]: audit 2026-03-21T06:51:30.629067+0000 mon.a (mon.0) 558 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:32.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:31 vm04 bash[20194]: cluster 2026-03-21T06:51:30.640278+0000 mon.a (mon.0) 559 : cluster [INF] osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646] boot 2026-03-21T06:51:32.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:31 vm04 bash[20194]: cluster 2026-03-21T06:51:30.640278+0000 mon.a (mon.0) 559 : cluster [INF] osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646] boot 2026-03-21T06:51:32.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:31 vm04 bash[20194]: cluster 2026-03-21T06:51:30.640313+0000 mon.a (mon.0) 560 : cluster [DBG] osdmap e36: 6 total, 6 up, 6 in 2026-03-21T06:51:32.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:31 vm04 bash[20194]: cluster 2026-03-21T06:51:30.640313+0000 mon.a (mon.0) 560 : cluster [DBG] osdmap e36: 6 total, 6 up, 6 in 2026-03-21T06:51:32.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:31 vm04 bash[20194]: audit 2026-03-21T06:51:30.641020+0000 mon.a (mon.0) 561 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:32.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:31 vm04 bash[20194]: audit 2026-03-21T06:51:30.641020+0000 mon.a (mon.0) 561 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:32.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:31 vm02 bash[17657]: cluster 2026-03-21T06:51:30.640278+0000 mon.a (mon.0) 559 : cluster [INF] osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646] boot 2026-03-21T06:51:32.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:31 vm02 bash[17657]: cluster 2026-03-21T06:51:30.640278+0000 mon.a (mon.0) 559 : cluster [INF] osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646] boot 2026-03-21T06:51:32.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:31 vm02 bash[17657]: cluster 2026-03-21T06:51:30.640313+0000 mon.a (mon.0) 560 : cluster [DBG] osdmap e36: 6 total, 6 up, 6 in 2026-03-21T06:51:32.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:31 vm02 bash[17657]: cluster 2026-03-21T06:51:30.640313+0000 mon.a (mon.0) 560 : cluster [DBG] osdmap e36: 6 total, 6 up, 6 in 2026-03-21T06:51:32.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:31 vm02 bash[17657]: audit 2026-03-21T06:51:30.641020+0000 mon.a (mon.0) 561 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:32.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:31 vm02 bash[17657]: audit 2026-03-21T06:51:30.641020+0000 mon.a (mon.0) 561 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:31 vm07 bash[19945]: cluster 2026-03-21T06:51:30.640278+0000 mon.a (mon.0) 559 : cluster [INF] osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646] boot 2026-03-21T06:51:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:31 vm07 bash[19945]: cluster 2026-03-21T06:51:30.640278+0000 mon.a (mon.0) 559 : cluster [INF] osd.5 [v2:192.168.123.107:6800/4080236646,v1:192.168.123.107:6801/4080236646] boot 2026-03-21T06:51:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:31 vm07 bash[19945]: cluster 2026-03-21T06:51:30.640313+0000 mon.a (mon.0) 560 : cluster [DBG] osdmap e36: 6 total, 6 up, 6 in 2026-03-21T06:51:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:31 vm07 bash[19945]: cluster 2026-03-21T06:51:30.640313+0000 mon.a (mon.0) 560 : cluster [DBG] osdmap e36: 6 total, 6 up, 6 in 2026-03-21T06:51:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:31 vm07 bash[19945]: audit 2026-03-21T06:51:30.641020+0000 mon.a (mon.0) 561 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:31 vm07 bash[19945]: audit 2026-03-21T06:51:30.641020+0000 mon.a (mon.0) 561 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:51:33.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:32 vm04 bash[20194]: cluster 2026-03-21T06:51:31.561913+0000 mgr.x (mgr.14152) 190 : cluster [DBG] pgmap v127: 1 pgs: 1 peering; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:33.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:32 vm04 bash[20194]: cluster 2026-03-21T06:51:31.561913+0000 mgr.x (mgr.14152) 190 : cluster [DBG] pgmap v127: 1 pgs: 1 peering; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:33.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:32 vm04 bash[20194]: cluster 2026-03-21T06:51:31.648107+0000 mon.a (mon.0) 562 : cluster [DBG] osdmap e37: 6 total, 6 up, 6 in 2026-03-21T06:51:33.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:32 vm04 bash[20194]: cluster 2026-03-21T06:51:31.648107+0000 mon.a (mon.0) 562 : cluster [DBG] osdmap e37: 6 total, 6 up, 6 in 2026-03-21T06:51:33.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:32 vm02 bash[17657]: cluster 2026-03-21T06:51:31.561913+0000 mgr.x (mgr.14152) 190 : cluster [DBG] pgmap v127: 1 pgs: 1 peering; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:33.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:32 vm02 bash[17657]: cluster 2026-03-21T06:51:31.561913+0000 mgr.x (mgr.14152) 190 : cluster [DBG] pgmap v127: 1 pgs: 1 peering; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:33.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:32 vm02 bash[17657]: cluster 2026-03-21T06:51:31.648107+0000 mon.a (mon.0) 562 : cluster [DBG] osdmap e37: 6 total, 6 up, 6 in 2026-03-21T06:51:33.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:32 vm02 bash[17657]: cluster 2026-03-21T06:51:31.648107+0000 mon.a (mon.0) 562 : cluster [DBG] osdmap e37: 6 total, 6 up, 6 in 2026-03-21T06:51:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:32 vm07 bash[19945]: cluster 2026-03-21T06:51:31.561913+0000 mgr.x (mgr.14152) 190 : cluster [DBG] pgmap v127: 1 pgs: 1 peering; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:32 vm07 bash[19945]: cluster 2026-03-21T06:51:31.561913+0000 mgr.x (mgr.14152) 190 : cluster [DBG] pgmap v127: 1 pgs: 1 peering; 449 KiB data, 135 MiB used, 100 GiB / 100 GiB avail 2026-03-21T06:51:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:32 vm07 bash[19945]: cluster 2026-03-21T06:51:31.648107+0000 mon.a (mon.0) 562 : cluster [DBG] osdmap e37: 6 total, 6 up, 6 in 2026-03-21T06:51:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:32 vm07 bash[19945]: cluster 2026-03-21T06:51:31.648107+0000 mon.a (mon.0) 562 : cluster [DBG] osdmap e37: 6 total, 6 up, 6 in 2026-03-21T06:51:34.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:33 vm04 bash[20194]: cluster 2026-03-21T06:51:32.651196+0000 mon.a (mon.0) 563 : cluster [DBG] osdmap e38: 6 total, 6 up, 6 in 2026-03-21T06:51:34.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:33 vm04 bash[20194]: cluster 2026-03-21T06:51:32.651196+0000 mon.a (mon.0) 563 : cluster [DBG] osdmap e38: 6 total, 6 up, 6 in 2026-03-21T06:51:34.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:33 vm02 bash[17657]: cluster 2026-03-21T06:51:32.651196+0000 mon.a (mon.0) 563 : cluster [DBG] osdmap e38: 6 total, 6 up, 6 in 2026-03-21T06:51:34.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:33 vm02 bash[17657]: cluster 2026-03-21T06:51:32.651196+0000 mon.a (mon.0) 563 : cluster [DBG] osdmap e38: 6 total, 6 up, 6 in 2026-03-21T06:51:34.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:33 vm07 bash[19945]: cluster 2026-03-21T06:51:32.651196+0000 mon.a (mon.0) 563 : cluster [DBG] osdmap e38: 6 total, 6 up, 6 in 2026-03-21T06:51:34.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:33 vm07 bash[19945]: cluster 2026-03-21T06:51:32.651196+0000 mon.a (mon.0) 563 : cluster [DBG] osdmap e38: 6 total, 6 up, 6 in 2026-03-21T06:51:35.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:34 vm04 bash[20194]: cluster 2026-03-21T06:51:33.562160+0000 mgr.x (mgr.14152) 191 : cluster [DBG] pgmap v130: 1 pgs: 1 peering; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:35.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:34 vm04 bash[20194]: cluster 2026-03-21T06:51:33.562160+0000 mgr.x (mgr.14152) 191 : cluster [DBG] pgmap v130: 1 pgs: 1 peering; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:35.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:34 vm02 bash[17657]: cluster 2026-03-21T06:51:33.562160+0000 mgr.x (mgr.14152) 191 : cluster [DBG] pgmap v130: 1 pgs: 1 peering; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:35.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:34 vm02 bash[17657]: cluster 2026-03-21T06:51:33.562160+0000 mgr.x (mgr.14152) 191 : cluster [DBG] pgmap v130: 1 pgs: 1 peering; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:35.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:34 vm07 bash[19945]: cluster 2026-03-21T06:51:33.562160+0000 mgr.x (mgr.14152) 191 : cluster [DBG] pgmap v130: 1 pgs: 1 peering; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:35.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:34 vm07 bash[19945]: cluster 2026-03-21T06:51:33.562160+0000 mgr.x (mgr.14152) 191 : cluster [DBG] pgmap v130: 1 pgs: 1 peering; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:37.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:36 vm04 bash[20194]: cluster 2026-03-21T06:51:35.562406+0000 mgr.x (mgr.14152) 192 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 76 KiB/s, 0 objects/s recovering 2026-03-21T06:51:37.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:36 vm04 bash[20194]: cluster 2026-03-21T06:51:35.562406+0000 mgr.x (mgr.14152) 192 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 76 KiB/s, 0 objects/s recovering 2026-03-21T06:51:37.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:36 vm02 bash[17657]: cluster 2026-03-21T06:51:35.562406+0000 mgr.x (mgr.14152) 192 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 76 KiB/s, 0 objects/s recovering 2026-03-21T06:51:37.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:36 vm02 bash[17657]: cluster 2026-03-21T06:51:35.562406+0000 mgr.x (mgr.14152) 192 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 76 KiB/s, 0 objects/s recovering 2026-03-21T06:51:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:36 vm07 bash[19945]: cluster 2026-03-21T06:51:35.562406+0000 mgr.x (mgr.14152) 192 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 76 KiB/s, 0 objects/s recovering 2026-03-21T06:51:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:36 vm07 bash[19945]: cluster 2026-03-21T06:51:35.562406+0000 mgr.x (mgr.14152) 192 : cluster [DBG] pgmap v131: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 76 KiB/s, 0 objects/s recovering 2026-03-21T06:51:39.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:38 vm04 bash[20194]: cluster 2026-03-21T06:51:37.562686+0000 mgr.x (mgr.14152) 193 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-21T06:51:39.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:38 vm04 bash[20194]: cluster 2026-03-21T06:51:37.562686+0000 mgr.x (mgr.14152) 193 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-21T06:51:39.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:38 vm02 bash[17657]: cluster 2026-03-21T06:51:37.562686+0000 mgr.x (mgr.14152) 193 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-21T06:51:39.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:38 vm02 bash[17657]: cluster 2026-03-21T06:51:37.562686+0000 mgr.x (mgr.14152) 193 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-21T06:51:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:38 vm07 bash[19945]: cluster 2026-03-21T06:51:37.562686+0000 mgr.x (mgr.14152) 193 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-21T06:51:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:38 vm07 bash[19945]: cluster 2026-03-21T06:51:37.562686+0000 mgr.x (mgr.14152) 193 : cluster [DBG] pgmap v132: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 65 KiB/s, 0 objects/s recovering 2026-03-21T06:51:41.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:40 vm04 bash[20194]: cluster 2026-03-21T06:51:39.562930+0000 mgr.x (mgr.14152) 194 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-21T06:51:41.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:40 vm04 bash[20194]: cluster 2026-03-21T06:51:39.562930+0000 mgr.x (mgr.14152) 194 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-21T06:51:41.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:40 vm02 bash[17657]: cluster 2026-03-21T06:51:39.562930+0000 mgr.x (mgr.14152) 194 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-21T06:51:41.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:40 vm02 bash[17657]: cluster 2026-03-21T06:51:39.562930+0000 mgr.x (mgr.14152) 194 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-21T06:51:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:40 vm07 bash[19945]: cluster 2026-03-21T06:51:39.562930+0000 mgr.x (mgr.14152) 194 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-21T06:51:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:40 vm07 bash[19945]: cluster 2026-03-21T06:51:39.562930+0000 mgr.x (mgr.14152) 194 : cluster [DBG] pgmap v133: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 56 KiB/s, 0 objects/s recovering 2026-03-21T06:51:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:42 vm04 bash[20194]: cluster 2026-03-21T06:51:41.563230+0000 mgr.x (mgr.14152) 195 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 45 KiB/s, 0 objects/s recovering 2026-03-21T06:51:43.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:42 vm04 bash[20194]: cluster 2026-03-21T06:51:41.563230+0000 mgr.x (mgr.14152) 195 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 45 KiB/s, 0 objects/s recovering 2026-03-21T06:51:43.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:42 vm02 bash[17657]: cluster 2026-03-21T06:51:41.563230+0000 mgr.x (mgr.14152) 195 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 45 KiB/s, 0 objects/s recovering 2026-03-21T06:51:43.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:42 vm02 bash[17657]: cluster 2026-03-21T06:51:41.563230+0000 mgr.x (mgr.14152) 195 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 45 KiB/s, 0 objects/s recovering 2026-03-21T06:51:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:42 vm07 bash[19945]: cluster 2026-03-21T06:51:41.563230+0000 mgr.x (mgr.14152) 195 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 45 KiB/s, 0 objects/s recovering 2026-03-21T06:51:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:42 vm07 bash[19945]: cluster 2026-03-21T06:51:41.563230+0000 mgr.x (mgr.14152) 195 : cluster [DBG] pgmap v134: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 45 KiB/s, 0 objects/s recovering 2026-03-21T06:51:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:43 vm04 bash[20194]: cluster 2026-03-21T06:51:43.563547+0000 mgr.x (mgr.14152) 196 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 41 KiB/s, 0 objects/s recovering 2026-03-21T06:51:44.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:43 vm04 bash[20194]: cluster 2026-03-21T06:51:43.563547+0000 mgr.x (mgr.14152) 196 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 41 KiB/s, 0 objects/s recovering 2026-03-21T06:51:44.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:43 vm02 bash[17657]: cluster 2026-03-21T06:51:43.563547+0000 mgr.x (mgr.14152) 196 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 41 KiB/s, 0 objects/s recovering 2026-03-21T06:51:44.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:43 vm02 bash[17657]: cluster 2026-03-21T06:51:43.563547+0000 mgr.x (mgr.14152) 196 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 41 KiB/s, 0 objects/s recovering 2026-03-21T06:51:44.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:43 vm07 bash[19945]: cluster 2026-03-21T06:51:43.563547+0000 mgr.x (mgr.14152) 196 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 41 KiB/s, 0 objects/s recovering 2026-03-21T06:51:44.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:43 vm07 bash[19945]: cluster 2026-03-21T06:51:43.563547+0000 mgr.x (mgr.14152) 196 : cluster [DBG] pgmap v135: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 41 KiB/s, 0 objects/s recovering 2026-03-21T06:51:46.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:46 vm02 bash[17657]: cluster 2026-03-21T06:51:45.563859+0000 mgr.x (mgr.14152) 197 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 37 KiB/s, 0 objects/s recovering 2026-03-21T06:51:46.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:46 vm02 bash[17657]: cluster 2026-03-21T06:51:45.563859+0000 mgr.x (mgr.14152) 197 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 37 KiB/s, 0 objects/s recovering 2026-03-21T06:51:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:46 vm07 bash[19945]: cluster 2026-03-21T06:51:45.563859+0000 mgr.x (mgr.14152) 197 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 37 KiB/s, 0 objects/s recovering 2026-03-21T06:51:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:46 vm07 bash[19945]: cluster 2026-03-21T06:51:45.563859+0000 mgr.x (mgr.14152) 197 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 37 KiB/s, 0 objects/s recovering 2026-03-21T06:51:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:46 vm04 bash[20194]: cluster 2026-03-21T06:51:45.563859+0000 mgr.x (mgr.14152) 197 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 37 KiB/s, 0 objects/s recovering 2026-03-21T06:51:47.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:46 vm04 bash[20194]: cluster 2026-03-21T06:51:45.563859+0000 mgr.x (mgr.14152) 197 : cluster [DBG] pgmap v136: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail; 37 KiB/s, 0 objects/s recovering 2026-03-21T06:51:48.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:48 vm02 bash[17657]: cluster 2026-03-21T06:51:47.564106+0000 mgr.x (mgr.14152) 198 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:48.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:48 vm02 bash[17657]: cluster 2026-03-21T06:51:47.564106+0000 mgr.x (mgr.14152) 198 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:48 vm07 bash[19945]: cluster 2026-03-21T06:51:47.564106+0000 mgr.x (mgr.14152) 198 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:48 vm07 bash[19945]: cluster 2026-03-21T06:51:47.564106+0000 mgr.x (mgr.14152) 198 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:49.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:48 vm04 bash[20194]: cluster 2026-03-21T06:51:47.564106+0000 mgr.x (mgr.14152) 198 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:49.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:48 vm04 bash[20194]: cluster 2026-03-21T06:51:47.564106+0000 mgr.x (mgr.14152) 198 : cluster [DBG] pgmap v137: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:50.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:50 vm02 bash[17657]: cluster 2026-03-21T06:51:49.564324+0000 mgr.x (mgr.14152) 199 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:50.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:50 vm02 bash[17657]: cluster 2026-03-21T06:51:49.564324+0000 mgr.x (mgr.14152) 199 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:50 vm07 bash[19945]: cluster 2026-03-21T06:51:49.564324+0000 mgr.x (mgr.14152) 199 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:50 vm07 bash[19945]: cluster 2026-03-21T06:51:49.564324+0000 mgr.x (mgr.14152) 199 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:51.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:50 vm04 bash[20194]: cluster 2026-03-21T06:51:49.564324+0000 mgr.x (mgr.14152) 199 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:51.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:50 vm04 bash[20194]: cluster 2026-03-21T06:51:49.564324+0000 mgr.x (mgr.14152) 199 : cluster [DBG] pgmap v138: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:52 vm02 bash[17657]: cluster 2026-03-21T06:51:51.564570+0000 mgr.x (mgr.14152) 200 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:52.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:52 vm02 bash[17657]: cluster 2026-03-21T06:51:51.564570+0000 mgr.x (mgr.14152) 200 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:52.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:52 vm07 bash[19945]: cluster 2026-03-21T06:51:51.564570+0000 mgr.x (mgr.14152) 200 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:52.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:52 vm07 bash[19945]: cluster 2026-03-21T06:51:51.564570+0000 mgr.x (mgr.14152) 200 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:53.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:52 vm04 bash[20194]: cluster 2026-03-21T06:51:51.564570+0000 mgr.x (mgr.14152) 200 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:53.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:52 vm04 bash[20194]: cluster 2026-03-21T06:51:51.564570+0000 mgr.x (mgr.14152) 200 : cluster [DBG] pgmap v139: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:54.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:54 vm07 bash[19945]: cluster 2026-03-21T06:51:53.564827+0000 mgr.x (mgr.14152) 201 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:54.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:54 vm07 bash[19945]: cluster 2026-03-21T06:51:53.564827+0000 mgr.x (mgr.14152) 201 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:55.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:54 vm04 bash[20194]: cluster 2026-03-21T06:51:53.564827+0000 mgr.x (mgr.14152) 201 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:55.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:54 vm04 bash[20194]: cluster 2026-03-21T06:51:53.564827+0000 mgr.x (mgr.14152) 201 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:55.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:54 vm02 bash[17657]: cluster 2026-03-21T06:51:53.564827+0000 mgr.x (mgr.14152) 201 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:55.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:54 vm02 bash[17657]: cluster 2026-03-21T06:51:53.564827+0000 mgr.x (mgr.14152) 201 : cluster [DBG] pgmap v140: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:57.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:56 vm04 bash[20194]: cluster 2026-03-21T06:51:55.565096+0000 mgr.x (mgr.14152) 202 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:57.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:56 vm04 bash[20194]: cluster 2026-03-21T06:51:55.565096+0000 mgr.x (mgr.14152) 202 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:57.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:56 vm02 bash[17657]: cluster 2026-03-21T06:51:55.565096+0000 mgr.x (mgr.14152) 202 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:57.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:56 vm02 bash[17657]: cluster 2026-03-21T06:51:55.565096+0000 mgr.x (mgr.14152) 202 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:57.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:56 vm07 bash[19945]: cluster 2026-03-21T06:51:55.565096+0000 mgr.x (mgr.14152) 202 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:57.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:56 vm07 bash[19945]: cluster 2026-03-21T06:51:55.565096+0000 mgr.x (mgr.14152) 202 : cluster [DBG] pgmap v141: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:59.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:58 vm04 bash[20194]: cluster 2026-03-21T06:51:57.565379+0000 mgr.x (mgr.14152) 203 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:59.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:51:58 vm04 bash[20194]: cluster 2026-03-21T06:51:57.565379+0000 mgr.x (mgr.14152) 203 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:59.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:58 vm02 bash[17657]: cluster 2026-03-21T06:51:57.565379+0000 mgr.x (mgr.14152) 203 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:59.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:51:58 vm02 bash[17657]: cluster 2026-03-21T06:51:57.565379+0000 mgr.x (mgr.14152) 203 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:58 vm07 bash[19945]: cluster 2026-03-21T06:51:57.565379+0000 mgr.x (mgr.14152) 203 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:51:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:51:58 vm07 bash[19945]: cluster 2026-03-21T06:51:57.565379+0000 mgr.x (mgr.14152) 203 : cluster [DBG] pgmap v142: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:01.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:00 vm04 bash[20194]: cluster 2026-03-21T06:51:59.565674+0000 mgr.x (mgr.14152) 204 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:01.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:00 vm04 bash[20194]: cluster 2026-03-21T06:51:59.565674+0000 mgr.x (mgr.14152) 204 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:00 vm02 bash[17657]: cluster 2026-03-21T06:51:59.565674+0000 mgr.x (mgr.14152) 204 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:01.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:00 vm02 bash[17657]: cluster 2026-03-21T06:51:59.565674+0000 mgr.x (mgr.14152) 204 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:01.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:00 vm07 bash[19945]: cluster 2026-03-21T06:51:59.565674+0000 mgr.x (mgr.14152) 204 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:01.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:00 vm07 bash[19945]: cluster 2026-03-21T06:51:59.565674+0000 mgr.x (mgr.14152) 204 : cluster [DBG] pgmap v143: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:03.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:02 vm04 bash[20194]: cluster 2026-03-21T06:52:01.565951+0000 mgr.x (mgr.14152) 205 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:03.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:02 vm04 bash[20194]: cluster 2026-03-21T06:52:01.565951+0000 mgr.x (mgr.14152) 205 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:03.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:02 vm02 bash[17657]: cluster 2026-03-21T06:52:01.565951+0000 mgr.x (mgr.14152) 205 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:03.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:02 vm02 bash[17657]: cluster 2026-03-21T06:52:01.565951+0000 mgr.x (mgr.14152) 205 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:03.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:02 vm07 bash[19945]: cluster 2026-03-21T06:52:01.565951+0000 mgr.x (mgr.14152) 205 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:03.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:02 vm07 bash[19945]: cluster 2026-03-21T06:52:01.565951+0000 mgr.x (mgr.14152) 205 : cluster [DBG] pgmap v144: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:05.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:04 vm04 bash[20194]: cluster 2026-03-21T06:52:03.566253+0000 mgr.x (mgr.14152) 206 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:05.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:04 vm04 bash[20194]: cluster 2026-03-21T06:52:03.566253+0000 mgr.x (mgr.14152) 206 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:05.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:04 vm02 bash[17657]: cluster 2026-03-21T06:52:03.566253+0000 mgr.x (mgr.14152) 206 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:05.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:04 vm02 bash[17657]: cluster 2026-03-21T06:52:03.566253+0000 mgr.x (mgr.14152) 206 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:05.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:04 vm07 bash[19945]: cluster 2026-03-21T06:52:03.566253+0000 mgr.x (mgr.14152) 206 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:05.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:04 vm07 bash[19945]: cluster 2026-03-21T06:52:03.566253+0000 mgr.x (mgr.14152) 206 : cluster [DBG] pgmap v145: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:06 vm04 bash[20194]: cluster 2026-03-21T06:52:05.566522+0000 mgr.x (mgr.14152) 207 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:07.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:06 vm04 bash[20194]: cluster 2026-03-21T06:52:05.566522+0000 mgr.x (mgr.14152) 207 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:07.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:06 vm02 bash[17657]: cluster 2026-03-21T06:52:05.566522+0000 mgr.x (mgr.14152) 207 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:07.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:06 vm02 bash[17657]: cluster 2026-03-21T06:52:05.566522+0000 mgr.x (mgr.14152) 207 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:06 vm07 bash[19945]: cluster 2026-03-21T06:52:05.566522+0000 mgr.x (mgr.14152) 207 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:06 vm07 bash[19945]: cluster 2026-03-21T06:52:05.566522+0000 mgr.x (mgr.14152) 207 : cluster [DBG] pgmap v146: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:09.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:08 vm04 bash[20194]: cluster 2026-03-21T06:52:07.566805+0000 mgr.x (mgr.14152) 208 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:09.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:08 vm04 bash[20194]: cluster 2026-03-21T06:52:07.566805+0000 mgr.x (mgr.14152) 208 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:08 vm02 bash[17657]: cluster 2026-03-21T06:52:07.566805+0000 mgr.x (mgr.14152) 208 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:09.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:08 vm02 bash[17657]: cluster 2026-03-21T06:52:07.566805+0000 mgr.x (mgr.14152) 208 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:08 vm07 bash[19945]: cluster 2026-03-21T06:52:07.566805+0000 mgr.x (mgr.14152) 208 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:08 vm07 bash[19945]: cluster 2026-03-21T06:52:07.566805+0000 mgr.x (mgr.14152) 208 : cluster [DBG] pgmap v147: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:10.003 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:09 vm04 bash[20194]: cluster 2026-03-21T06:52:09.567120+0000 mgr.x (mgr.14152) 209 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:10.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:09 vm04 bash[20194]: cluster 2026-03-21T06:52:09.567120+0000 mgr.x (mgr.14152) 209 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:10.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:09 vm02 bash[17657]: cluster 2026-03-21T06:52:09.567120+0000 mgr.x (mgr.14152) 209 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:10.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:09 vm02 bash[17657]: cluster 2026-03-21T06:52:09.567120+0000 mgr.x (mgr.14152) 209 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:10.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:09 vm07 bash[19945]: cluster 2026-03-21T06:52:09.567120+0000 mgr.x (mgr.14152) 209 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:10.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:09 vm07 bash[19945]: cluster 2026-03-21T06:52:09.567120+0000 mgr.x (mgr.14152) 209 : cluster [DBG] pgmap v148: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:12.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:12 vm02 bash[17657]: cluster 2026-03-21T06:52:11.567463+0000 mgr.x (mgr.14152) 210 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:12.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:12 vm02 bash[17657]: cluster 2026-03-21T06:52:11.567463+0000 mgr.x (mgr.14152) 210 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:12.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:12 vm07 bash[19945]: cluster 2026-03-21T06:52:11.567463+0000 mgr.x (mgr.14152) 210 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:12.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:12 vm07 bash[19945]: cluster 2026-03-21T06:52:11.567463+0000 mgr.x (mgr.14152) 210 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:13.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:12 vm04 bash[20194]: cluster 2026-03-21T06:52:11.567463+0000 mgr.x (mgr.14152) 210 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:13.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:12 vm04 bash[20194]: cluster 2026-03-21T06:52:11.567463+0000 mgr.x (mgr.14152) 210 : cluster [DBG] pgmap v149: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:14.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:14 vm02 bash[17657]: cluster 2026-03-21T06:52:13.567784+0000 mgr.x (mgr.14152) 211 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:14.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:14 vm02 bash[17657]: cluster 2026-03-21T06:52:13.567784+0000 mgr.x (mgr.14152) 211 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:14.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:14 vm07 bash[19945]: cluster 2026-03-21T06:52:13.567784+0000 mgr.x (mgr.14152) 211 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:14.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:14 vm07 bash[19945]: cluster 2026-03-21T06:52:13.567784+0000 mgr.x (mgr.14152) 211 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:15.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:14 vm04 bash[20194]: cluster 2026-03-21T06:52:13.567784+0000 mgr.x (mgr.14152) 211 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:15.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:14 vm04 bash[20194]: cluster 2026-03-21T06:52:13.567784+0000 mgr.x (mgr.14152) 211 : cluster [DBG] pgmap v150: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:16 vm02 bash[17657]: cluster 2026-03-21T06:52:15.568088+0000 mgr.x (mgr.14152) 212 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:16.896 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:16 vm02 bash[17657]: cluster 2026-03-21T06:52:15.568088+0000 mgr.x (mgr.14152) 212 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:16.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:16 vm07 bash[19945]: cluster 2026-03-21T06:52:15.568088+0000 mgr.x (mgr.14152) 212 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:16.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:16 vm07 bash[19945]: cluster 2026-03-21T06:52:15.568088+0000 mgr.x (mgr.14152) 212 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:17.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:16 vm04 bash[20194]: cluster 2026-03-21T06:52:15.568088+0000 mgr.x (mgr.14152) 212 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:17.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:16 vm04 bash[20194]: cluster 2026-03-21T06:52:15.568088+0000 mgr.x (mgr.14152) 212 : cluster [DBG] pgmap v151: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:18.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:18 vm07 bash[19945]: cluster 2026-03-21T06:52:17.568377+0000 mgr.x (mgr.14152) 213 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:18.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:18 vm07 bash[19945]: cluster 2026-03-21T06:52:17.568377+0000 mgr.x (mgr.14152) 213 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:19.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:18 vm04 bash[20194]: cluster 2026-03-21T06:52:17.568377+0000 mgr.x (mgr.14152) 213 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:19.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:18 vm04 bash[20194]: cluster 2026-03-21T06:52:17.568377+0000 mgr.x (mgr.14152) 213 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:18 vm02 bash[17657]: cluster 2026-03-21T06:52:17.568377+0000 mgr.x (mgr.14152) 213 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:19.146 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:18 vm02 bash[17657]: cluster 2026-03-21T06:52:17.568377+0000 mgr.x (mgr.14152) 213 : cluster [DBG] pgmap v152: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:20 vm04 bash[20194]: cluster 2026-03-21T06:52:19.568700+0000 mgr.x (mgr.14152) 214 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:21.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:20 vm04 bash[20194]: cluster 2026-03-21T06:52:19.568700+0000 mgr.x (mgr.14152) 214 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:21.145 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:20 vm02 bash[17657]: cluster 2026-03-21T06:52:19.568700+0000 mgr.x (mgr.14152) 214 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:21.145 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:20 vm02 bash[17657]: cluster 2026-03-21T06:52:19.568700+0000 mgr.x (mgr.14152) 214 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:20 vm07 bash[19945]: cluster 2026-03-21T06:52:19.568700+0000 mgr.x (mgr.14152) 214 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:20 vm07 bash[19945]: cluster 2026-03-21T06:52:19.568700+0000 mgr.x (mgr.14152) 214 : cluster [DBG] pgmap v153: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:23.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:22 vm04 bash[20194]: cluster 2026-03-21T06:52:21.569043+0000 mgr.x (mgr.14152) 215 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:23.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:22 vm04 bash[20194]: cluster 2026-03-21T06:52:21.569043+0000 mgr.x (mgr.14152) 215 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:23.145 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:22 vm02 bash[17657]: cluster 2026-03-21T06:52:21.569043+0000 mgr.x (mgr.14152) 215 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:23.145 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:22 vm02 bash[17657]: cluster 2026-03-21T06:52:21.569043+0000 mgr.x (mgr.14152) 215 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:22 vm07 bash[19945]: cluster 2026-03-21T06:52:21.569043+0000 mgr.x (mgr.14152) 215 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:22 vm07 bash[19945]: cluster 2026-03-21T06:52:21.569043+0000 mgr.x (mgr.14152) 215 : cluster [DBG] pgmap v154: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:24 vm04 bash[20194]: cluster 2026-03-21T06:52:23.569301+0000 mgr.x (mgr.14152) 216 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:25.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:24 vm04 bash[20194]: cluster 2026-03-21T06:52:23.569301+0000 mgr.x (mgr.14152) 216 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:25.145 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:24 vm02 bash[17657]: cluster 2026-03-21T06:52:23.569301+0000 mgr.x (mgr.14152) 216 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:25.145 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:24 vm02 bash[17657]: cluster 2026-03-21T06:52:23.569301+0000 mgr.x (mgr.14152) 216 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:25.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:24 vm07 bash[19945]: cluster 2026-03-21T06:52:23.569301+0000 mgr.x (mgr.14152) 216 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:25.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:24 vm07 bash[19945]: cluster 2026-03-21T06:52:23.569301+0000 mgr.x (mgr.14152) 216 : cluster [DBG] pgmap v155: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:27.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:26 vm04 bash[20194]: cluster 2026-03-21T06:52:25.569586+0000 mgr.x (mgr.14152) 217 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:27.004 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:26 vm04 bash[20194]: cluster 2026-03-21T06:52:25.569586+0000 mgr.x (mgr.14152) 217 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:27.144 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:26 vm02 bash[17657]: cluster 2026-03-21T06:52:25.569586+0000 mgr.x (mgr.14152) 217 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:27.144 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:26 vm02 bash[17657]: cluster 2026-03-21T06:52:25.569586+0000 mgr.x (mgr.14152) 217 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:26 vm07 bash[19945]: cluster 2026-03-21T06:52:25.569586+0000 mgr.x (mgr.14152) 217 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:26 vm07 bash[19945]: cluster 2026-03-21T06:52:25.569586+0000 mgr.x (mgr.14152) 217 : cluster [DBG] pgmap v156: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:29.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:28 vm04 bash[20194]: cluster 2026-03-21T06:52:27.569907+0000 mgr.x (mgr.14152) 218 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:29.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:28 vm04 bash[20194]: cluster 2026-03-21T06:52:27.569907+0000 mgr.x (mgr.14152) 218 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:29.144 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:28 vm02 bash[17657]: cluster 2026-03-21T06:52:27.569907+0000 mgr.x (mgr.14152) 218 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:29.144 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:28 vm02 bash[17657]: cluster 2026-03-21T06:52:27.569907+0000 mgr.x (mgr.14152) 218 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:28 vm07 bash[19945]: cluster 2026-03-21T06:52:27.569907+0000 mgr.x (mgr.14152) 218 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:28 vm07 bash[19945]: cluster 2026-03-21T06:52:27.569907+0000 mgr.x (mgr.14152) 218 : cluster [DBG] pgmap v157: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:31.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:30 vm04 bash[20194]: cluster 2026-03-21T06:52:29.570220+0000 mgr.x (mgr.14152) 219 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:31.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:30 vm04 bash[20194]: cluster 2026-03-21T06:52:29.570220+0000 mgr.x (mgr.14152) 219 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:31.144 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:30 vm02 bash[17657]: cluster 2026-03-21T06:52:29.570220+0000 mgr.x (mgr.14152) 219 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:31.144 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:30 vm02 bash[17657]: cluster 2026-03-21T06:52:29.570220+0000 mgr.x (mgr.14152) 219 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:30 vm07 bash[19945]: cluster 2026-03-21T06:52:29.570220+0000 mgr.x (mgr.14152) 219 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:30 vm07 bash[19945]: cluster 2026-03-21T06:52:29.570220+0000 mgr.x (mgr.14152) 219 : cluster [DBG] pgmap v158: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:33.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:32 vm04 bash[20194]: cluster 2026-03-21T06:52:31.570529+0000 mgr.x (mgr.14152) 220 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:33.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:32 vm04 bash[20194]: cluster 2026-03-21T06:52:31.570529+0000 mgr.x (mgr.14152) 220 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:33.144 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:32 vm02 bash[17657]: cluster 2026-03-21T06:52:31.570529+0000 mgr.x (mgr.14152) 220 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:33.144 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:32 vm02 bash[17657]: cluster 2026-03-21T06:52:31.570529+0000 mgr.x (mgr.14152) 220 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:32 vm07 bash[19945]: cluster 2026-03-21T06:52:31.570529+0000 mgr.x (mgr.14152) 220 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:32 vm07 bash[19945]: cluster 2026-03-21T06:52:31.570529+0000 mgr.x (mgr.14152) 220 : cluster [DBG] pgmap v159: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:35.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:34 vm04 bash[20194]: cluster 2026-03-21T06:52:33.570820+0000 mgr.x (mgr.14152) 221 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:35.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:34 vm04 bash[20194]: cluster 2026-03-21T06:52:33.570820+0000 mgr.x (mgr.14152) 221 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:35.144 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:34 vm02 bash[17657]: cluster 2026-03-21T06:52:33.570820+0000 mgr.x (mgr.14152) 221 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:35.144 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:34 vm02 bash[17657]: cluster 2026-03-21T06:52:33.570820+0000 mgr.x (mgr.14152) 221 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:35.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:34 vm07 bash[19945]: cluster 2026-03-21T06:52:33.570820+0000 mgr.x (mgr.14152) 221 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:35.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:34 vm07 bash[19945]: cluster 2026-03-21T06:52:33.570820+0000 mgr.x (mgr.14152) 221 : cluster [DBG] pgmap v160: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:37.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:36 vm04 bash[20194]: cluster 2026-03-21T06:52:35.571131+0000 mgr.x (mgr.14152) 222 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:37.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:36 vm04 bash[20194]: cluster 2026-03-21T06:52:35.571131+0000 mgr.x (mgr.14152) 222 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:37.143 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:36 vm02 bash[17657]: cluster 2026-03-21T06:52:35.571131+0000 mgr.x (mgr.14152) 222 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:37.143 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:36 vm02 bash[17657]: cluster 2026-03-21T06:52:35.571131+0000 mgr.x (mgr.14152) 222 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:36 vm07 bash[19945]: cluster 2026-03-21T06:52:35.571131+0000 mgr.x (mgr.14152) 222 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:36 vm07 bash[19945]: cluster 2026-03-21T06:52:35.571131+0000 mgr.x (mgr.14152) 222 : cluster [DBG] pgmap v161: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:38.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:37 vm04 bash[20194]: cluster 2026-03-21T06:52:37.571434+0000 mgr.x (mgr.14152) 223 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:38.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:37 vm04 bash[20194]: cluster 2026-03-21T06:52:37.571434+0000 mgr.x (mgr.14152) 223 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:38.143 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:37 vm02 bash[17657]: cluster 2026-03-21T06:52:37.571434+0000 mgr.x (mgr.14152) 223 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:38.143 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:37 vm02 bash[17657]: cluster 2026-03-21T06:52:37.571434+0000 mgr.x (mgr.14152) 223 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:38.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:37 vm07 bash[19945]: cluster 2026-03-21T06:52:37.571434+0000 mgr.x (mgr.14152) 223 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:38.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:37 vm07 bash[19945]: cluster 2026-03-21T06:52:37.571434+0000 mgr.x (mgr.14152) 223 : cluster [DBG] pgmap v162: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:40.893 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:40 vm02 bash[17657]: cluster 2026-03-21T06:52:39.571728+0000 mgr.x (mgr.14152) 224 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:40.893 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:40 vm02 bash[17657]: cluster 2026-03-21T06:52:39.571728+0000 mgr.x (mgr.14152) 224 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:40.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:40 vm07 bash[19945]: cluster 2026-03-21T06:52:39.571728+0000 mgr.x (mgr.14152) 224 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:40.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:40 vm07 bash[19945]: cluster 2026-03-21T06:52:39.571728+0000 mgr.x (mgr.14152) 224 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:41.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:40 vm04 bash[20194]: cluster 2026-03-21T06:52:39.571728+0000 mgr.x (mgr.14152) 224 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:41.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:40 vm04 bash[20194]: cluster 2026-03-21T06:52:39.571728+0000 mgr.x (mgr.14152) 224 : cluster [DBG] pgmap v163: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:42.893 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:42 vm02 bash[17657]: cluster 2026-03-21T06:52:41.572089+0000 mgr.x (mgr.14152) 225 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:42.893 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:42 vm02 bash[17657]: cluster 2026-03-21T06:52:41.572089+0000 mgr.x (mgr.14152) 225 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:42 vm07 bash[19945]: cluster 2026-03-21T06:52:41.572089+0000 mgr.x (mgr.14152) 225 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:42 vm07 bash[19945]: cluster 2026-03-21T06:52:41.572089+0000 mgr.x (mgr.14152) 225 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:43.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:42 vm04 bash[20194]: cluster 2026-03-21T06:52:41.572089+0000 mgr.x (mgr.14152) 225 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:43.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:42 vm04 bash[20194]: cluster 2026-03-21T06:52:41.572089+0000 mgr.x (mgr.14152) 225 : cluster [DBG] pgmap v164: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:44.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:44 vm07 bash[19945]: cluster 2026-03-21T06:52:43.572397+0000 mgr.x (mgr.14152) 226 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:44.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:44 vm07 bash[19945]: cluster 2026-03-21T06:52:43.572397+0000 mgr.x (mgr.14152) 226 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:45.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:44 vm04 bash[20194]: cluster 2026-03-21T06:52:43.572397+0000 mgr.x (mgr.14152) 226 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:45.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:44 vm04 bash[20194]: cluster 2026-03-21T06:52:43.572397+0000 mgr.x (mgr.14152) 226 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:45.143 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:44 vm02 bash[17657]: cluster 2026-03-21T06:52:43.572397+0000 mgr.x (mgr.14152) 226 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:45.143 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:44 vm02 bash[17657]: cluster 2026-03-21T06:52:43.572397+0000 mgr.x (mgr.14152) 226 : cluster [DBG] pgmap v165: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:46 vm07 bash[19945]: cluster 2026-03-21T06:52:45.572766+0000 mgr.x (mgr.14152) 227 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:46 vm07 bash[19945]: cluster 2026-03-21T06:52:45.572766+0000 mgr.x (mgr.14152) 227 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:47.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:46 vm04 bash[20194]: cluster 2026-03-21T06:52:45.572766+0000 mgr.x (mgr.14152) 227 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:47.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:46 vm04 bash[20194]: cluster 2026-03-21T06:52:45.572766+0000 mgr.x (mgr.14152) 227 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:47.143 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:46 vm02 bash[17657]: cluster 2026-03-21T06:52:45.572766+0000 mgr.x (mgr.14152) 227 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:47.143 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:46 vm02 bash[17657]: cluster 2026-03-21T06:52:45.572766+0000 mgr.x (mgr.14152) 227 : cluster [DBG] pgmap v166: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:49.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:48 vm04 bash[20194]: cluster 2026-03-21T06:52:47.573008+0000 mgr.x (mgr.14152) 228 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:49.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:48 vm04 bash[20194]: cluster 2026-03-21T06:52:47.573008+0000 mgr.x (mgr.14152) 228 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:49.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:48 vm02 bash[17657]: cluster 2026-03-21T06:52:47.573008+0000 mgr.x (mgr.14152) 228 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:49.143 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:48 vm02 bash[17657]: cluster 2026-03-21T06:52:47.573008+0000 mgr.x (mgr.14152) 228 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:49.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:48 vm07 bash[19945]: cluster 2026-03-21T06:52:47.573008+0000 mgr.x (mgr.14152) 228 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:49.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:48 vm07 bash[19945]: cluster 2026-03-21T06:52:47.573008+0000 mgr.x (mgr.14152) 228 : cluster [DBG] pgmap v167: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:51.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:50 vm04 bash[20194]: cluster 2026-03-21T06:52:49.573331+0000 mgr.x (mgr.14152) 229 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:51.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:50 vm04 bash[20194]: cluster 2026-03-21T06:52:49.573331+0000 mgr.x (mgr.14152) 229 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:51.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:50 vm02 bash[17657]: cluster 2026-03-21T06:52:49.573331+0000 mgr.x (mgr.14152) 229 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:51.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:50 vm02 bash[17657]: cluster 2026-03-21T06:52:49.573331+0000 mgr.x (mgr.14152) 229 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:50 vm07 bash[19945]: cluster 2026-03-21T06:52:49.573331+0000 mgr.x (mgr.14152) 229 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:50 vm07 bash[19945]: cluster 2026-03-21T06:52:49.573331+0000 mgr.x (mgr.14152) 229 : cluster [DBG] pgmap v168: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:53.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:52 vm04 bash[20194]: cluster 2026-03-21T06:52:51.573663+0000 mgr.x (mgr.14152) 230 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:53.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:52 vm04 bash[20194]: cluster 2026-03-21T06:52:51.573663+0000 mgr.x (mgr.14152) 230 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:53.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:52 vm02 bash[17657]: cluster 2026-03-21T06:52:51.573663+0000 mgr.x (mgr.14152) 230 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:53.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:52 vm02 bash[17657]: cluster 2026-03-21T06:52:51.573663+0000 mgr.x (mgr.14152) 230 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:52 vm07 bash[19945]: cluster 2026-03-21T06:52:51.573663+0000 mgr.x (mgr.14152) 230 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:52 vm07 bash[19945]: cluster 2026-03-21T06:52:51.573663+0000 mgr.x (mgr.14152) 230 : cluster [DBG] pgmap v169: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:55.005 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:54 vm04 bash[20194]: cluster 2026-03-21T06:52:53.573959+0000 mgr.x (mgr.14152) 231 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:55.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:54 vm04 bash[20194]: cluster 2026-03-21T06:52:53.573959+0000 mgr.x (mgr.14152) 231 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:55.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:54 vm02 bash[17657]: cluster 2026-03-21T06:52:53.573959+0000 mgr.x (mgr.14152) 231 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:55.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:54 vm02 bash[17657]: cluster 2026-03-21T06:52:53.573959+0000 mgr.x (mgr.14152) 231 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:55.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:54 vm07 bash[19945]: cluster 2026-03-21T06:52:53.573959+0000 mgr.x (mgr.14152) 231 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:55.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:54 vm07 bash[19945]: cluster 2026-03-21T06:52:53.573959+0000 mgr.x (mgr.14152) 231 : cluster [DBG] pgmap v170: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:57.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:56 vm04 bash[20194]: cluster 2026-03-21T06:52:55.574226+0000 mgr.x (mgr.14152) 232 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:57.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:56 vm04 bash[20194]: cluster 2026-03-21T06:52:55.574226+0000 mgr.x (mgr.14152) 232 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:57.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:56 vm02 bash[17657]: cluster 2026-03-21T06:52:55.574226+0000 mgr.x (mgr.14152) 232 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:57.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:56 vm02 bash[17657]: cluster 2026-03-21T06:52:55.574226+0000 mgr.x (mgr.14152) 232 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:57.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:56 vm07 bash[19945]: cluster 2026-03-21T06:52:55.574226+0000 mgr.x (mgr.14152) 232 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:57.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:56 vm07 bash[19945]: cluster 2026-03-21T06:52:55.574226+0000 mgr.x (mgr.14152) 232 : cluster [DBG] pgmap v171: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:59.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:58 vm04 bash[20194]: cluster 2026-03-21T06:52:57.574510+0000 mgr.x (mgr.14152) 233 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:59.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:52:58 vm04 bash[20194]: cluster 2026-03-21T06:52:57.574510+0000 mgr.x (mgr.14152) 233 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:59.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:58 vm02 bash[17657]: cluster 2026-03-21T06:52:57.574510+0000 mgr.x (mgr.14152) 233 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:59.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:52:58 vm02 bash[17657]: cluster 2026-03-21T06:52:57.574510+0000 mgr.x (mgr.14152) 233 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:58 vm07 bash[19945]: cluster 2026-03-21T06:52:57.574510+0000 mgr.x (mgr.14152) 233 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:52:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:52:58 vm07 bash[19945]: cluster 2026-03-21T06:52:57.574510+0000 mgr.x (mgr.14152) 233 : cluster [DBG] pgmap v172: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:01.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:00 vm04 bash[20194]: cluster 2026-03-21T06:52:59.574803+0000 mgr.x (mgr.14152) 234 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:01.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:00 vm04 bash[20194]: cluster 2026-03-21T06:52:59.574803+0000 mgr.x (mgr.14152) 234 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:01.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:00 vm02 bash[17657]: cluster 2026-03-21T06:52:59.574803+0000 mgr.x (mgr.14152) 234 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:01.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:00 vm02 bash[17657]: cluster 2026-03-21T06:52:59.574803+0000 mgr.x (mgr.14152) 234 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:01.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:00 vm07 bash[19945]: cluster 2026-03-21T06:52:59.574803+0000 mgr.x (mgr.14152) 234 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:01.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:00 vm07 bash[19945]: cluster 2026-03-21T06:52:59.574803+0000 mgr.x (mgr.14152) 234 : cluster [DBG] pgmap v173: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:03.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:02 vm04 bash[20194]: cluster 2026-03-21T06:53:01.575150+0000 mgr.x (mgr.14152) 235 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:03.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:02 vm04 bash[20194]: cluster 2026-03-21T06:53:01.575150+0000 mgr.x (mgr.14152) 235 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:03.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:02 vm02 bash[17657]: cluster 2026-03-21T06:53:01.575150+0000 mgr.x (mgr.14152) 235 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:03.142 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:02 vm02 bash[17657]: cluster 2026-03-21T06:53:01.575150+0000 mgr.x (mgr.14152) 235 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:03.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:02 vm07 bash[19945]: cluster 2026-03-21T06:53:01.575150+0000 mgr.x (mgr.14152) 235 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:03.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:02 vm07 bash[19945]: cluster 2026-03-21T06:53:01.575150+0000 mgr.x (mgr.14152) 235 : cluster [DBG] pgmap v174: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:04.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:03 vm04 bash[20194]: cluster 2026-03-21T06:53:03.575429+0000 mgr.x (mgr.14152) 236 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:04.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:03 vm04 bash[20194]: cluster 2026-03-21T06:53:03.575429+0000 mgr.x (mgr.14152) 236 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:04.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:03 vm02 bash[17657]: cluster 2026-03-21T06:53:03.575429+0000 mgr.x (mgr.14152) 236 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:04.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:03 vm02 bash[17657]: cluster 2026-03-21T06:53:03.575429+0000 mgr.x (mgr.14152) 236 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:04.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:03 vm07 bash[19945]: cluster 2026-03-21T06:53:03.575429+0000 mgr.x (mgr.14152) 236 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:04.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:03 vm07 bash[19945]: cluster 2026-03-21T06:53:03.575429+0000 mgr.x (mgr.14152) 236 : cluster [DBG] pgmap v175: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:06.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:06 vm02 bash[17657]: cluster 2026-03-21T06:53:05.575680+0000 mgr.x (mgr.14152) 237 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:06.892 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:06 vm02 bash[17657]: cluster 2026-03-21T06:53:05.575680+0000 mgr.x (mgr.14152) 237 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:06 vm07 bash[19945]: cluster 2026-03-21T06:53:05.575680+0000 mgr.x (mgr.14152) 237 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:06 vm07 bash[19945]: cluster 2026-03-21T06:53:05.575680+0000 mgr.x (mgr.14152) 237 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:07.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:06 vm04 bash[20194]: cluster 2026-03-21T06:53:05.575680+0000 mgr.x (mgr.14152) 237 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:07.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:06 vm04 bash[20194]: cluster 2026-03-21T06:53:05.575680+0000 mgr.x (mgr.14152) 237 : cluster [DBG] pgmap v176: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:08.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:08 vm02 bash[17657]: cluster 2026-03-21T06:53:07.575963+0000 mgr.x (mgr.14152) 238 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:08.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:08 vm02 bash[17657]: cluster 2026-03-21T06:53:07.575963+0000 mgr.x (mgr.14152) 238 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:08.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:08 vm07 bash[19945]: cluster 2026-03-21T06:53:07.575963+0000 mgr.x (mgr.14152) 238 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:08.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:08 vm07 bash[19945]: cluster 2026-03-21T06:53:07.575963+0000 mgr.x (mgr.14152) 238 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:09.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:08 vm04 bash[20194]: cluster 2026-03-21T06:53:07.575963+0000 mgr.x (mgr.14152) 238 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:09.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:08 vm04 bash[20194]: cluster 2026-03-21T06:53:07.575963+0000 mgr.x (mgr.14152) 238 : cluster [DBG] pgmap v177: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:10.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:10 vm07 bash[19945]: cluster 2026-03-21T06:53:09.576205+0000 mgr.x (mgr.14152) 239 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:10.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:10 vm07 bash[19945]: cluster 2026-03-21T06:53:09.576205+0000 mgr.x (mgr.14152) 239 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:11.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:10 vm04 bash[20194]: cluster 2026-03-21T06:53:09.576205+0000 mgr.x (mgr.14152) 239 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:11.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:10 vm04 bash[20194]: cluster 2026-03-21T06:53:09.576205+0000 mgr.x (mgr.14152) 239 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:11.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:10 vm02 bash[17657]: cluster 2026-03-21T06:53:09.576205+0000 mgr.x (mgr.14152) 239 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:11.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:10 vm02 bash[17657]: cluster 2026-03-21T06:53:09.576205+0000 mgr.x (mgr.14152) 239 : cluster [DBG] pgmap v178: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:12.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:12 vm07 bash[19945]: cluster 2026-03-21T06:53:11.576476+0000 mgr.x (mgr.14152) 240 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:12.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:12 vm07 bash[19945]: cluster 2026-03-21T06:53:11.576476+0000 mgr.x (mgr.14152) 240 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:13.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:12 vm04 bash[20194]: cluster 2026-03-21T06:53:11.576476+0000 mgr.x (mgr.14152) 240 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:13.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:12 vm04 bash[20194]: cluster 2026-03-21T06:53:11.576476+0000 mgr.x (mgr.14152) 240 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:13.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:12 vm02 bash[17657]: cluster 2026-03-21T06:53:11.576476+0000 mgr.x (mgr.14152) 240 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:13.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:12 vm02 bash[17657]: cluster 2026-03-21T06:53:11.576476+0000 mgr.x (mgr.14152) 240 : cluster [DBG] pgmap v179: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:15.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:14 vm04 bash[20194]: cluster 2026-03-21T06:53:13.577075+0000 mgr.x (mgr.14152) 241 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:15.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:14 vm04 bash[20194]: cluster 2026-03-21T06:53:13.577075+0000 mgr.x (mgr.14152) 241 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:15.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:14 vm02 bash[17657]: cluster 2026-03-21T06:53:13.577075+0000 mgr.x (mgr.14152) 241 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:15.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:14 vm02 bash[17657]: cluster 2026-03-21T06:53:13.577075+0000 mgr.x (mgr.14152) 241 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:14 vm07 bash[19945]: cluster 2026-03-21T06:53:13.577075+0000 mgr.x (mgr.14152) 241 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:14 vm07 bash[19945]: cluster 2026-03-21T06:53:13.577075+0000 mgr.x (mgr.14152) 241 : cluster [DBG] pgmap v180: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:17.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:16 vm04 bash[20194]: cluster 2026-03-21T06:53:15.577457+0000 mgr.x (mgr.14152) 242 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:17.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:16 vm04 bash[20194]: cluster 2026-03-21T06:53:15.577457+0000 mgr.x (mgr.14152) 242 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:17.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:16 vm02 bash[17657]: cluster 2026-03-21T06:53:15.577457+0000 mgr.x (mgr.14152) 242 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:17.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:16 vm02 bash[17657]: cluster 2026-03-21T06:53:15.577457+0000 mgr.x (mgr.14152) 242 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:16 vm07 bash[19945]: cluster 2026-03-21T06:53:15.577457+0000 mgr.x (mgr.14152) 242 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:16 vm07 bash[19945]: cluster 2026-03-21T06:53:15.577457+0000 mgr.x (mgr.14152) 242 : cluster [DBG] pgmap v181: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:19.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:18 vm04 bash[20194]: cluster 2026-03-21T06:53:17.577756+0000 mgr.x (mgr.14152) 243 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:19.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:18 vm04 bash[20194]: cluster 2026-03-21T06:53:17.577756+0000 mgr.x (mgr.14152) 243 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:19.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:18 vm02 bash[17657]: cluster 2026-03-21T06:53:17.577756+0000 mgr.x (mgr.14152) 243 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:19.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:18 vm02 bash[17657]: cluster 2026-03-21T06:53:17.577756+0000 mgr.x (mgr.14152) 243 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:18 vm07 bash[19945]: cluster 2026-03-21T06:53:17.577756+0000 mgr.x (mgr.14152) 243 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:18 vm07 bash[19945]: cluster 2026-03-21T06:53:17.577756+0000 mgr.x (mgr.14152) 243 : cluster [DBG] pgmap v182: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:21.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:20 vm04 bash[20194]: cluster 2026-03-21T06:53:19.578087+0000 mgr.x (mgr.14152) 244 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:21.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:20 vm04 bash[20194]: cluster 2026-03-21T06:53:19.578087+0000 mgr.x (mgr.14152) 244 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:21.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:20 vm02 bash[17657]: cluster 2026-03-21T06:53:19.578087+0000 mgr.x (mgr.14152) 244 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:21.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:20 vm02 bash[17657]: cluster 2026-03-21T06:53:19.578087+0000 mgr.x (mgr.14152) 244 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:20 vm07 bash[19945]: cluster 2026-03-21T06:53:19.578087+0000 mgr.x (mgr.14152) 244 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:20 vm07 bash[19945]: cluster 2026-03-21T06:53:19.578087+0000 mgr.x (mgr.14152) 244 : cluster [DBG] pgmap v183: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:23.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:22 vm04 bash[20194]: cluster 2026-03-21T06:53:21.578413+0000 mgr.x (mgr.14152) 245 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:23.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:22 vm04 bash[20194]: cluster 2026-03-21T06:53:21.578413+0000 mgr.x (mgr.14152) 245 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:23.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:22 vm02 bash[17657]: cluster 2026-03-21T06:53:21.578413+0000 mgr.x (mgr.14152) 245 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:23.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:22 vm02 bash[17657]: cluster 2026-03-21T06:53:21.578413+0000 mgr.x (mgr.14152) 245 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:22 vm07 bash[19945]: cluster 2026-03-21T06:53:21.578413+0000 mgr.x (mgr.14152) 245 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:22 vm07 bash[19945]: cluster 2026-03-21T06:53:21.578413+0000 mgr.x (mgr.14152) 245 : cluster [DBG] pgmap v184: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:25.006 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:24 vm04 bash[20194]: cluster 2026-03-21T06:53:23.578730+0000 mgr.x (mgr.14152) 246 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:25.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:24 vm04 bash[20194]: cluster 2026-03-21T06:53:23.578730+0000 mgr.x (mgr.14152) 246 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:25.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:24 vm02 bash[17657]: cluster 2026-03-21T06:53:23.578730+0000 mgr.x (mgr.14152) 246 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:25.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:24 vm02 bash[17657]: cluster 2026-03-21T06:53:23.578730+0000 mgr.x (mgr.14152) 246 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:25.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:24 vm07 bash[19945]: cluster 2026-03-21T06:53:23.578730+0000 mgr.x (mgr.14152) 246 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:25.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:24 vm07 bash[19945]: cluster 2026-03-21T06:53:23.578730+0000 mgr.x (mgr.14152) 246 : cluster [DBG] pgmap v185: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:26.412 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:26 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:27.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:26 vm04 bash[20194]: cluster 2026-03-21T06:53:25.578984+0000 mgr.x (mgr.14152) 247 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:27.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:26 vm04 bash[20194]: cluster 2026-03-21T06:53:25.578984+0000 mgr.x (mgr.14152) 247 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:27.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:26 vm04 bash[20194]: audit 2026-03-21T06:53:26.345903+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:26 vm04 bash[20194]: audit 2026-03-21T06:53:26.345903+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:26 vm04 bash[20194]: audit 2026-03-21T06:53:26.350136+0000 mon.a (mon.0) 565 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:26 vm04 bash[20194]: audit 2026-03-21T06:53:26.350136+0000 mon.a (mon.0) 565 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:26 vm02 bash[17657]: cluster 2026-03-21T06:53:25.578984+0000 mgr.x (mgr.14152) 247 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:27.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:26 vm02 bash[17657]: cluster 2026-03-21T06:53:25.578984+0000 mgr.x (mgr.14152) 247 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:27.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:26 vm02 bash[17657]: audit 2026-03-21T06:53:26.345903+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:26 vm02 bash[17657]: audit 2026-03-21T06:53:26.345903+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:26 vm02 bash[17657]: audit 2026-03-21T06:53:26.350136+0000 mon.a (mon.0) 565 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:26 vm02 bash[17657]: audit 2026-03-21T06:53:26.350136+0000 mon.a (mon.0) 565 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:26 vm07 bash[19945]: cluster 2026-03-21T06:53:25.578984+0000 mgr.x (mgr.14152) 247 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:26 vm07 bash[19945]: cluster 2026-03-21T06:53:25.578984+0000 mgr.x (mgr.14152) 247 : cluster [DBG] pgmap v186: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:26 vm07 bash[19945]: audit 2026-03-21T06:53:26.345903+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:26 vm07 bash[19945]: audit 2026-03-21T06:53:26.345903+0000 mon.a (mon.0) 564 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:26 vm07 bash[19945]: audit 2026-03-21T06:53:26.350136+0000 mon.a (mon.0) 565 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:26 vm07 bash[19945]: audit 2026-03-21T06:53:26.350136+0000 mon.a (mon.0) 565 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:27.772 INFO:teuthology.orchestra.run.vm07.stdout:Created osd(s) 5 on host 'vm07' 2026-03-21T06:53:27.892 DEBUG:teuthology.orchestra.run.vm07:osd.5> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.5.service 2026-03-21T06:53:27.893 INFO:tasks.cephadm:Deploying osd.6 on vm07 with /dev/vg_nvme/lv_3... 2026-03-21T06:53:27.893 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- lvm zap /dev/vg_nvme/lv_3 2026-03-21T06:53:28.026 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:27 vm07 bash[24245]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-5 2026-03-21T06:53:28.027 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:27 vm07 bash[24245]: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/vg_nvme/lv_4 --path /var/lib/ceph/osd/ceph-5 --no-mon-config 2026-03-21T06:53:28.027 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:28 vm07 bash[24245]: Running command: /usr/bin/ln -snf /dev/vg_nvme/lv_4 /var/lib/ceph/osd/ceph-5/block 2026-03-21T06:53:28.151 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:28 vm07 bash[24245]: Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-5/block 2026-03-21T06:53:28.151 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:28 vm07 bash[24245]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-3 2026-03-21T06:53:28.151 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:28 vm07 bash[24245]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-5 2026-03-21T06:53:28.151 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:28 vm07 bash[24245]: --> ceph-volume lvm activate successful for osd ID: 5 2026-03-21T06:53:28.194 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:53:28.299 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:28 vm07 bash[25087]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-5 2026-03-21T06:53:28.988 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:28 vm07 bash[25201]: debug 2026-03-21T06:53:28.726+0000 7f8a49b1e8c0 -1 Falling back to public interface 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:27.579425+0000 mgr.x (mgr.14152) 248 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:27.579425+0000 mgr.x (mgr.14152) 248 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: audit 2026-03-21T06:53:27.762802+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: audit 2026-03-21T06:53:27.762802+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: audit 2026-03-21T06:53:27.766610+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: audit 2026-03-21T06:53:27.766610+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.396599+0000 mon.a (mon.0) 568 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.396599+0000 mon.a (mon.0) 568 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.396704+0000 mon.a (mon.0) 569 : cluster [INF] osd.5 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.396704+0000 mon.a (mon.0) 569 : cluster [INF] osd.5 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.396814+0000 mon.a (mon.0) 570 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.396814+0000 mon.a (mon.0) 570 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.396994+0000 mon.a (mon.0) 571 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.396994+0000 mon.a (mon.0) 571 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.397034+0000 mon.a (mon.0) 572 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.397034+0000 mon.a (mon.0) 572 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.397643+0000 mon.a (mon.0) 573 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.397643+0000 mon.a (mon.0) 573 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.397734+0000 mon.a (mon.0) 574 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.397734+0000 mon.a (mon.0) 574 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.397896+0000 mon.a (mon.0) 575 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.397896+0000 mon.a (mon.0) 575 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.397978+0000 mon.a (mon.0) 576 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.397978+0000 mon.a (mon.0) 576 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.398107+0000 mon.a (mon.0) 577 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.398107+0000 mon.a (mon.0) 577 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.398185+0000 mon.a (mon.0) 578 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:28.988 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:28 vm07 bash[19945]: cluster 2026-03-21T06:53:28.398185+0000 mon.a (mon.0) 578 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:27.579425+0000 mgr.x (mgr.14152) 248 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:27.579425+0000 mgr.x (mgr.14152) 248 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: audit 2026-03-21T06:53:27.762802+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: audit 2026-03-21T06:53:27.762802+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: audit 2026-03-21T06:53:27.766610+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: audit 2026-03-21T06:53:27.766610+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.396599+0000 mon.a (mon.0) 568 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.396599+0000 mon.a (mon.0) 568 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.396704+0000 mon.a (mon.0) 569 : cluster [INF] osd.5 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.396704+0000 mon.a (mon.0) 569 : cluster [INF] osd.5 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.396814+0000 mon.a (mon.0) 570 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.396814+0000 mon.a (mon.0) 570 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.396994+0000 mon.a (mon.0) 571 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.396994+0000 mon.a (mon.0) 571 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.397034+0000 mon.a (mon.0) 572 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.397034+0000 mon.a (mon.0) 572 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.397643+0000 mon.a (mon.0) 573 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.397643+0000 mon.a (mon.0) 573 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.397734+0000 mon.a (mon.0) 574 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.397734+0000 mon.a (mon.0) 574 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.397896+0000 mon.a (mon.0) 575 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.397896+0000 mon.a (mon.0) 575 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.397978+0000 mon.a (mon.0) 576 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.397978+0000 mon.a (mon.0) 576 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.398107+0000 mon.a (mon.0) 577 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.398107+0000 mon.a (mon.0) 577 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.398185+0000 mon.a (mon.0) 578 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:29.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:28 vm04 bash[20194]: cluster 2026-03-21T06:53:28.398185+0000 mon.a (mon.0) 578 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:27.579425+0000 mgr.x (mgr.14152) 248 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:27.579425+0000 mgr.x (mgr.14152) 248 : cluster [DBG] pgmap v187: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: audit 2026-03-21T06:53:27.762802+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: audit 2026-03-21T06:53:27.762802+0000 mon.a (mon.0) 566 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: audit 2026-03-21T06:53:27.766610+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: audit 2026-03-21T06:53:27.766610+0000 mon.a (mon.0) 567 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.396599+0000 mon.a (mon.0) 568 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.396599+0000 mon.a (mon.0) 568 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.396704+0000 mon.a (mon.0) 569 : cluster [INF] osd.5 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.396704+0000 mon.a (mon.0) 569 : cluster [INF] osd.5 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.396814+0000 mon.a (mon.0) 570 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.396814+0000 mon.a (mon.0) 570 : cluster [DBG] osd.5 reported immediately failed by osd.0 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.396994+0000 mon.a (mon.0) 571 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.396994+0000 mon.a (mon.0) 571 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.397034+0000 mon.a (mon.0) 572 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.397034+0000 mon.a (mon.0) 572 : cluster [DBG] osd.5 reported immediately failed by osd.1 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.397643+0000 mon.a (mon.0) 573 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.397643+0000 mon.a (mon.0) 573 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.397734+0000 mon.a (mon.0) 574 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.397734+0000 mon.a (mon.0) 574 : cluster [DBG] osd.5 reported immediately failed by osd.4 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.397896+0000 mon.a (mon.0) 575 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.397896+0000 mon.a (mon.0) 575 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.397978+0000 mon.a (mon.0) 576 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.397978+0000 mon.a (mon.0) 576 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.398107+0000 mon.a (mon.0) 577 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.398107+0000 mon.a (mon.0) 577 : cluster [DBG] osd.5 reported immediately failed by osd.2 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.398185+0000 mon.a (mon.0) 578 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:29.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:28 vm02 bash[17657]: cluster 2026-03-21T06:53:28.398185+0000 mon.a (mon.0) 578 : cluster [DBG] osd.5 reported immediately failed by osd.3 2026-03-21T06:53:29.401 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:29 vm07 bash[25201]: debug 2026-03-21T06:53:29.014+0000 7f8a49b1e8c0 -1 osd.5 38 log_to_monitors true 2026-03-21T06:53:29.522 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:53:29.538 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch daemon add osd vm07:vg_nvme/lv_3 --skip-validation 2026-03-21T06:53:29.774 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: cluster 2026-03-21T06:53:28.765918+0000 mon.a (mon.0) 579 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: cluster 2026-03-21T06:53:28.765918+0000 mon.a (mon.0) 579 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: cluster 2026-03-21T06:53:28.765948+0000 mon.a (mon.0) 580 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: cluster 2026-03-21T06:53:28.765948+0000 mon.a (mon.0) 580 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: cluster 2026-03-21T06:53:28.770052+0000 mon.a (mon.0) 581 : cluster [DBG] osdmap e39: 6 total, 5 up, 6 in 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: cluster 2026-03-21T06:53:28.770052+0000 mon.a (mon.0) 581 : cluster [DBG] osdmap e39: 6 total, 5 up, 6 in 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: audit 2026-03-21T06:53:29.027524+0000 mon.c (mon.1) 12 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: audit 2026-03-21T06:53:29.027524+0000 mon.c (mon.1) 12 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: audit 2026-03-21T06:53:29.027819+0000 mon.a (mon.0) 582 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: audit 2026-03-21T06:53:29.027819+0000 mon.a (mon.0) 582 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: audit 2026-03-21T06:53:29.044156+0000 mon.a (mon.0) 583 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: audit 2026-03-21T06:53:29.044156+0000 mon.a (mon.0) 583 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: audit 2026-03-21T06:53:29.048422+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: audit 2026-03-21T06:53:29.048422+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: audit 2026-03-21T06:53:29.062977+0000 mon.a (mon.0) 585 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: audit 2026-03-21T06:53:29.062977+0000 mon.a (mon.0) 585 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: cluster 2026-03-21T06:53:29.579766+0000 mgr.x (mgr.14152) 249 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:29.788 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:29 vm07 bash[19945]: cluster 2026-03-21T06:53:29.579766+0000 mgr.x (mgr.14152) 249 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: cluster 2026-03-21T06:53:28.765918+0000 mon.a (mon.0) 579 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: cluster 2026-03-21T06:53:28.765918+0000 mon.a (mon.0) 579 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: cluster 2026-03-21T06:53:28.765948+0000 mon.a (mon.0) 580 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: cluster 2026-03-21T06:53:28.765948+0000 mon.a (mon.0) 580 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: cluster 2026-03-21T06:53:28.770052+0000 mon.a (mon.0) 581 : cluster [DBG] osdmap e39: 6 total, 5 up, 6 in 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: cluster 2026-03-21T06:53:28.770052+0000 mon.a (mon.0) 581 : cluster [DBG] osdmap e39: 6 total, 5 up, 6 in 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: audit 2026-03-21T06:53:29.027524+0000 mon.c (mon.1) 12 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: audit 2026-03-21T06:53:29.027524+0000 mon.c (mon.1) 12 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: audit 2026-03-21T06:53:29.027819+0000 mon.a (mon.0) 582 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: audit 2026-03-21T06:53:29.027819+0000 mon.a (mon.0) 582 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: audit 2026-03-21T06:53:29.044156+0000 mon.a (mon.0) 583 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: audit 2026-03-21T06:53:29.044156+0000 mon.a (mon.0) 583 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: audit 2026-03-21T06:53:29.048422+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: audit 2026-03-21T06:53:29.048422+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: audit 2026-03-21T06:53:29.062977+0000 mon.a (mon.0) 585 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: audit 2026-03-21T06:53:29.062977+0000 mon.a (mon.0) 585 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: cluster 2026-03-21T06:53:29.579766+0000 mgr.x (mgr.14152) 249 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:30.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:29 vm04 bash[20194]: cluster 2026-03-21T06:53:29.579766+0000 mgr.x (mgr.14152) 249 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: cluster 2026-03-21T06:53:28.765918+0000 mon.a (mon.0) 579 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: cluster 2026-03-21T06:53:28.765918+0000 mon.a (mon.0) 579 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: cluster 2026-03-21T06:53:28.765948+0000 mon.a (mon.0) 580 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: cluster 2026-03-21T06:53:28.765948+0000 mon.a (mon.0) 580 : cluster [WRN] Health check failed: 1 host (1 osds) down (OSD_HOST_DOWN) 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: cluster 2026-03-21T06:53:28.770052+0000 mon.a (mon.0) 581 : cluster [DBG] osdmap e39: 6 total, 5 up, 6 in 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: cluster 2026-03-21T06:53:28.770052+0000 mon.a (mon.0) 581 : cluster [DBG] osdmap e39: 6 total, 5 up, 6 in 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: audit 2026-03-21T06:53:29.027524+0000 mon.c (mon.1) 12 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: audit 2026-03-21T06:53:29.027524+0000 mon.c (mon.1) 12 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: audit 2026-03-21T06:53:29.027819+0000 mon.a (mon.0) 582 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: audit 2026-03-21T06:53:29.027819+0000 mon.a (mon.0) 582 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]} : dispatch 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: audit 2026-03-21T06:53:29.044156+0000 mon.a (mon.0) 583 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: audit 2026-03-21T06:53:29.044156+0000 mon.a (mon.0) 583 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: audit 2026-03-21T06:53:29.048422+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: audit 2026-03-21T06:53:29.048422+0000 mon.a (mon.0) 584 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: audit 2026-03-21T06:53:29.062977+0000 mon.a (mon.0) 585 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: audit 2026-03-21T06:53:29.062977+0000 mon.a (mon.0) 585 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: cluster 2026-03-21T06:53:29.579766+0000 mgr.x (mgr.14152) 249 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:30.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:29 vm02 bash[17657]: cluster 2026-03-21T06:53:29.579766+0000 mgr.x (mgr.14152) 249 : cluster [DBG] pgmap v189: 1 pgs: 1 active+clean; 449 KiB data, 161 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:30.151 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:29 vm07 bash[25201]: debug 2026-03-21T06:53:29.794+0000 7f8a41086640 -1 osd.5 38 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:29.771260+0000 mon.a (mon.0) 586 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:29.771260+0000 mon.a (mon.0) 586 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: cluster 2026-03-21T06:53:29.773609+0000 mon.a (mon.0) 587 : cluster [DBG] osdmap e40: 6 total, 5 up, 6 in 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: cluster 2026-03-21T06:53:29.773609+0000 mon.a (mon.0) 587 : cluster [DBG] osdmap e40: 6 total, 5 up, 6 in 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:29.774694+0000 mon.c (mon.1) 13 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:29.774694+0000 mon.c (mon.1) 13 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:29.774885+0000 mon.a (mon.0) 588 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:29.774885+0000 mon.a (mon.0) 588 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.117270+0000 mgr.x (mgr.14152) 250 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.117270+0000 mgr.x (mgr.14152) 250 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: cephadm 2026-03-21T06:53:30.118052+0000 mgr.x (mgr.14152) 251 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: cephadm 2026-03-21T06:53:30.118052+0000 mgr.x (mgr.14152) 251 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: cephadm 2026-03-21T06:53:30.118120+0000 mgr.x (mgr.14152) 252 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_3'] 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: cephadm 2026-03-21T06:53:30.118120+0000 mgr.x (mgr.14152) 252 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_3'] 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.122239+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.122239+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: cephadm 2026-03-21T06:53:30.122545+0000 mgr.x (mgr.14152) 253 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: cephadm 2026-03-21T06:53:30.122545+0000 mgr.x (mgr.14152) 253 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: cephadm 2026-03-21T06:53:30.122644+0000 mgr.x (mgr.14152) 254 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: cephadm 2026-03-21T06:53:30.122644+0000 mgr.x (mgr.14152) 254 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.125440+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.125440+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.126232+0000 mon.a (mon.0) 591 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.126232+0000 mon.a (mon.0) 591 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.127886+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.127886+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.128371+0000 mon.a (mon.0) 593 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:31.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:30 vm02 bash[17657]: audit 2026-03-21T06:53:30.128371+0000 mon.a (mon.0) 593 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:29.771260+0000 mon.a (mon.0) 586 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:29.771260+0000 mon.a (mon.0) 586 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: cluster 2026-03-21T06:53:29.773609+0000 mon.a (mon.0) 587 : cluster [DBG] osdmap e40: 6 total, 5 up, 6 in 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: cluster 2026-03-21T06:53:29.773609+0000 mon.a (mon.0) 587 : cluster [DBG] osdmap e40: 6 total, 5 up, 6 in 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:29.774694+0000 mon.c (mon.1) 13 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:29.774694+0000 mon.c (mon.1) 13 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:29.774885+0000 mon.a (mon.0) 588 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:29.774885+0000 mon.a (mon.0) 588 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.117270+0000 mgr.x (mgr.14152) 250 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.117270+0000 mgr.x (mgr.14152) 250 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: cephadm 2026-03-21T06:53:30.118052+0000 mgr.x (mgr.14152) 251 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: cephadm 2026-03-21T06:53:30.118052+0000 mgr.x (mgr.14152) 251 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: cephadm 2026-03-21T06:53:30.118120+0000 mgr.x (mgr.14152) 252 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_3'] 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: cephadm 2026-03-21T06:53:30.118120+0000 mgr.x (mgr.14152) 252 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_3'] 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.122239+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.122239+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: cephadm 2026-03-21T06:53:30.122545+0000 mgr.x (mgr.14152) 253 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: cephadm 2026-03-21T06:53:30.122545+0000 mgr.x (mgr.14152) 253 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: cephadm 2026-03-21T06:53:30.122644+0000 mgr.x (mgr.14152) 254 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: cephadm 2026-03-21T06:53:30.122644+0000 mgr.x (mgr.14152) 254 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.125440+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.125440+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.126232+0000 mon.a (mon.0) 591 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.126232+0000 mon.a (mon.0) 591 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.127886+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.127886+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.128371+0000 mon.a (mon.0) 593 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:30 vm07 bash[19945]: audit 2026-03-21T06:53:30.128371+0000 mon.a (mon.0) 593 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:29.771260+0000 mon.a (mon.0) 586 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:29.771260+0000 mon.a (mon.0) 586 : audit [INF] from='osd.5 ' entity='osd.5' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["5"]}]': finished 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: cluster 2026-03-21T06:53:29.773609+0000 mon.a (mon.0) 587 : cluster [DBG] osdmap e40: 6 total, 5 up, 6 in 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: cluster 2026-03-21T06:53:29.773609+0000 mon.a (mon.0) 587 : cluster [DBG] osdmap e40: 6 total, 5 up, 6 in 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:29.774694+0000 mon.c (mon.1) 13 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:29.774694+0000 mon.c (mon.1) 13 : audit [INF] from='osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835]' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:29.774885+0000 mon.a (mon.0) 588 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:29.774885+0000 mon.a (mon.0) 588 : audit [INF] from='osd.5 ' entity='osd.5' cmd={"prefix": "osd crush create-or-move", "id": 5, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.117270+0000 mgr.x (mgr.14152) 250 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.117270+0000 mgr.x (mgr.14152) 250 : audit [DBG] from='client.24274 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_3", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: cephadm 2026-03-21T06:53:30.118052+0000 mgr.x (mgr.14152) 251 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: cephadm 2026-03-21T06:53:30.118052+0000 mgr.x (mgr.14152) 251 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: cephadm 2026-03-21T06:53:30.118120+0000 mgr.x (mgr.14152) 252 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_3'] 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: cephadm 2026-03-21T06:53:30.118120+0000 mgr.x (mgr.14152) 252 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_3'] 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.122239+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.122239+0000 mon.a (mon.0) 589 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: cephadm 2026-03-21T06:53:30.122545+0000 mgr.x (mgr.14152) 253 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:31.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: cephadm 2026-03-21T06:53:30.122545+0000 mgr.x (mgr.14152) 253 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:31.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: cephadm 2026-03-21T06:53:30.122644+0000 mgr.x (mgr.14152) 254 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:31.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: cephadm 2026-03-21T06:53:30.122644+0000 mgr.x (mgr.14152) 254 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:31.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.125440+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.125440+0000 mon.a (mon.0) 590 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:31.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.126232+0000 mon.a (mon.0) 591 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:31.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.126232+0000 mon.a (mon.0) 591 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:31.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.127886+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:31.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.127886+0000 mon.a (mon.0) 592 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:31.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.128371+0000 mon.a (mon.0) 593 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:31.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:30 vm04 bash[20194]: audit 2026-03-21T06:53:30.128371+0000 mon.a (mon.0) 593 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:30.772538+0000 mon.a (mon.0) 594 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:30.772538+0000 mon.a (mon.0) 594 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:30.772550+0000 mon.a (mon.0) 595 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:30.772550+0000 mon.a (mon.0) 595 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:30.772555+0000 mon.a (mon.0) 596 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:30.772555+0000 mon.a (mon.0) 596 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:30.787099+0000 mon.a (mon.0) 597 : cluster [INF] osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835] boot 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:30.787099+0000 mon.a (mon.0) 597 : cluster [INF] osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835] boot 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:30.787138+0000 mon.a (mon.0) 598 : cluster [DBG] osdmap e41: 6 total, 6 up, 6 in 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:30.787138+0000 mon.a (mon.0) 598 : cluster [DBG] osdmap e41: 6 total, 6 up, 6 in 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:30.787284+0000 mon.a (mon.0) 599 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:30.787284+0000 mon.a (mon.0) 599 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:31.011510+0000 mon.c (mon.1) 14 : audit [INF] from='client.? 192.168.123.107:0/3162770629' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:31.011510+0000 mon.c (mon.1) 14 : audit [INF] from='client.? 192.168.123.107:0/3162770629' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:31.011707+0000 mon.a (mon.0) 600 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:31.011707+0000 mon.a (mon.0) 600 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:31.015206+0000 mon.a (mon.0) 601 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"}]': finished 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:31.015206+0000 mon.a (mon.0) 601 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"}]': finished 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:31.016994+0000 mon.a (mon.0) 602 : cluster [DBG] osdmap e42: 7 total, 6 up, 7 in 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:31.016994+0000 mon.a (mon.0) 602 : cluster [DBG] osdmap e42: 7 total, 6 up, 7 in 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:31.017120+0000 mon.a (mon.0) 603 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:31.017120+0000 mon.a (mon.0) 603 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:31.382281+0000 mon.c (mon.1) 15 : audit [DBG] from='client.? 192.168.123.107:0/2225087575' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: audit 2026-03-21T06:53:31.382281+0000 mon.c (mon.1) 15 : audit [DBG] from='client.? 192.168.123.107:0/2225087575' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:31.580053+0000 mgr.x (mgr.14152) 255 : cluster [DBG] pgmap v193: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:32.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:31 vm02 bash[17657]: cluster 2026-03-21T06:53:31.580053+0000 mgr.x (mgr.14152) 255 : cluster [DBG] pgmap v193: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:30.772538+0000 mon.a (mon.0) 594 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:30.772538+0000 mon.a (mon.0) 594 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:30.772550+0000 mon.a (mon.0) 595 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:30.772550+0000 mon.a (mon.0) 595 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:30.772555+0000 mon.a (mon.0) 596 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:30.772555+0000 mon.a (mon.0) 596 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:30.787099+0000 mon.a (mon.0) 597 : cluster [INF] osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835] boot 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:30.787099+0000 mon.a (mon.0) 597 : cluster [INF] osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835] boot 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:30.787138+0000 mon.a (mon.0) 598 : cluster [DBG] osdmap e41: 6 total, 6 up, 6 in 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:30.787138+0000 mon.a (mon.0) 598 : cluster [DBG] osdmap e41: 6 total, 6 up, 6 in 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:30.787284+0000 mon.a (mon.0) 599 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:30.787284+0000 mon.a (mon.0) 599 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:31.011510+0000 mon.c (mon.1) 14 : audit [INF] from='client.? 192.168.123.107:0/3162770629' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:31.011510+0000 mon.c (mon.1) 14 : audit [INF] from='client.? 192.168.123.107:0/3162770629' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:31.011707+0000 mon.a (mon.0) 600 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:31.011707+0000 mon.a (mon.0) 600 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:31.015206+0000 mon.a (mon.0) 601 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"}]': finished 2026-03-21T06:53:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:31.015206+0000 mon.a (mon.0) 601 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"}]': finished 2026-03-21T06:53:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:31.016994+0000 mon.a (mon.0) 602 : cluster [DBG] osdmap e42: 7 total, 6 up, 7 in 2026-03-21T06:53:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:31.016994+0000 mon.a (mon.0) 602 : cluster [DBG] osdmap e42: 7 total, 6 up, 7 in 2026-03-21T06:53:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:31.017120+0000 mon.a (mon.0) 603 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:31.017120+0000 mon.a (mon.0) 603 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:31.382281+0000 mon.c (mon.1) 15 : audit [DBG] from='client.? 192.168.123.107:0/2225087575' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: audit 2026-03-21T06:53:31.382281+0000 mon.c (mon.1) 15 : audit [DBG] from='client.? 192.168.123.107:0/2225087575' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:31.580053+0000 mgr.x (mgr.14152) 255 : cluster [DBG] pgmap v193: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:31 vm07 bash[19945]: cluster 2026-03-21T06:53:31.580053+0000 mgr.x (mgr.14152) 255 : cluster [DBG] pgmap v193: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:30.772538+0000 mon.a (mon.0) 594 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:30.772538+0000 mon.a (mon.0) 594 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:30.772550+0000 mon.a (mon.0) 595 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:30.772550+0000 mon.a (mon.0) 595 : cluster [INF] Health check cleared: OSD_HOST_DOWN (was: 1 host (1 osds) down) 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:30.772555+0000 mon.a (mon.0) 596 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:30.772555+0000 mon.a (mon.0) 596 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:30.787099+0000 mon.a (mon.0) 597 : cluster [INF] osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835] boot 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:30.787099+0000 mon.a (mon.0) 597 : cluster [INF] osd.5 [v2:192.168.123.107:6800/102330835,v1:192.168.123.107:6801/102330835] boot 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:30.787138+0000 mon.a (mon.0) 598 : cluster [DBG] osdmap e41: 6 total, 6 up, 6 in 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:30.787138+0000 mon.a (mon.0) 598 : cluster [DBG] osdmap e41: 6 total, 6 up, 6 in 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:30.787284+0000 mon.a (mon.0) 599 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:30.787284+0000 mon.a (mon.0) 599 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 5} : dispatch 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:31.011510+0000 mon.c (mon.1) 14 : audit [INF] from='client.? 192.168.123.107:0/3162770629' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:31.011510+0000 mon.c (mon.1) 14 : audit [INF] from='client.? 192.168.123.107:0/3162770629' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:31.011707+0000 mon.a (mon.0) 600 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:31.011707+0000 mon.a (mon.0) 600 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"} : dispatch 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:31.015206+0000 mon.a (mon.0) 601 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"}]': finished 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:31.015206+0000 mon.a (mon.0) 601 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be"}]': finished 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:31.016994+0000 mon.a (mon.0) 602 : cluster [DBG] osdmap e42: 7 total, 6 up, 7 in 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:31.016994+0000 mon.a (mon.0) 602 : cluster [DBG] osdmap e42: 7 total, 6 up, 7 in 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:31.017120+0000 mon.a (mon.0) 603 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:31.017120+0000 mon.a (mon.0) 603 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:31.382281+0000 mon.c (mon.1) 15 : audit [DBG] from='client.? 192.168.123.107:0/2225087575' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: audit 2026-03-21T06:53:31.382281+0000 mon.c (mon.1) 15 : audit [DBG] from='client.? 192.168.123.107:0/2225087575' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:31.580053+0000 mgr.x (mgr.14152) 255 : cluster [DBG] pgmap v193: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:32.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:31 vm04 bash[20194]: cluster 2026-03-21T06:53:31.580053+0000 mgr.x (mgr.14152) 255 : cluster [DBG] pgmap v193: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:33.058 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:32 vm07 bash[19945]: cluster 2026-03-21T06:53:32.015264+0000 mon.a (mon.0) 604 : cluster [WRN] Health check failed: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded (PG_DEGRADED) 2026-03-21T06:53:33.058 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:32 vm07 bash[19945]: cluster 2026-03-21T06:53:32.015264+0000 mon.a (mon.0) 604 : cluster [WRN] Health check failed: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded (PG_DEGRADED) 2026-03-21T06:53:33.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:32 vm02 bash[17657]: cluster 2026-03-21T06:53:32.015264+0000 mon.a (mon.0) 604 : cluster [WRN] Health check failed: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded (PG_DEGRADED) 2026-03-21T06:53:33.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:32 vm02 bash[17657]: cluster 2026-03-21T06:53:32.015264+0000 mon.a (mon.0) 604 : cluster [WRN] Health check failed: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded (PG_DEGRADED) 2026-03-21T06:53:33.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:32 vm04 bash[20194]: cluster 2026-03-21T06:53:32.015264+0000 mon.a (mon.0) 604 : cluster [WRN] Health check failed: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded (PG_DEGRADED) 2026-03-21T06:53:33.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:32 vm04 bash[20194]: cluster 2026-03-21T06:53:32.015264+0000 mon.a (mon.0) 604 : cluster [WRN] Health check failed: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded (PG_DEGRADED) 2026-03-21T06:53:33.916 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:33 vm07 bash[19945]: audit 2026-03-21T06:53:33.099519+0000 mon.a (mon.0) 605 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:33.916 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:33 vm07 bash[19945]: audit 2026-03-21T06:53:33.099519+0000 mon.a (mon.0) 605 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:33.916 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:33 vm07 bash[19945]: audit 2026-03-21T06:53:33.100036+0000 mon.a (mon.0) 606 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:33.916 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:33 vm07 bash[19945]: audit 2026-03-21T06:53:33.100036+0000 mon.a (mon.0) 606 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:33.916 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:33 vm07 bash[19945]: cephadm 2026-03-21T06:53:33.100441+0000 mgr.x (mgr.14152) 256 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:33.916 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:33 vm07 bash[19945]: cephadm 2026-03-21T06:53:33.100441+0000 mgr.x (mgr.14152) 256 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:33.916 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:33 vm07 bash[19945]: cluster 2026-03-21T06:53:33.580325+0000 mgr.x (mgr.14152) 257 : cluster [DBG] pgmap v194: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:33.916 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:33 vm07 bash[19945]: cluster 2026-03-21T06:53:33.580325+0000 mgr.x (mgr.14152) 257 : cluster [DBG] pgmap v194: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:33.916 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:33 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:33.916 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:33 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:34.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:33 vm02 bash[17657]: audit 2026-03-21T06:53:33.099519+0000 mon.a (mon.0) 605 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:34.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:33 vm02 bash[17657]: audit 2026-03-21T06:53:33.099519+0000 mon.a (mon.0) 605 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:34.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:33 vm02 bash[17657]: audit 2026-03-21T06:53:33.100036+0000 mon.a (mon.0) 606 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:34.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:33 vm02 bash[17657]: audit 2026-03-21T06:53:33.100036+0000 mon.a (mon.0) 606 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:34.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:33 vm02 bash[17657]: cephadm 2026-03-21T06:53:33.100441+0000 mgr.x (mgr.14152) 256 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:34.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:33 vm02 bash[17657]: cephadm 2026-03-21T06:53:33.100441+0000 mgr.x (mgr.14152) 256 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:34.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:33 vm02 bash[17657]: cluster 2026-03-21T06:53:33.580325+0000 mgr.x (mgr.14152) 257 : cluster [DBG] pgmap v194: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:34.141 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:33 vm02 bash[17657]: cluster 2026-03-21T06:53:33.580325+0000 mgr.x (mgr.14152) 257 : cluster [DBG] pgmap v194: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:34.173 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:34 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:34.173 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:34 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:34.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:33 vm04 bash[20194]: audit 2026-03-21T06:53:33.099519+0000 mon.a (mon.0) 605 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:34.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:33 vm04 bash[20194]: audit 2026-03-21T06:53:33.099519+0000 mon.a (mon.0) 605 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:34.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:33 vm04 bash[20194]: audit 2026-03-21T06:53:33.100036+0000 mon.a (mon.0) 606 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:34.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:33 vm04 bash[20194]: audit 2026-03-21T06:53:33.100036+0000 mon.a (mon.0) 606 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:34.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:33 vm04 bash[20194]: cephadm 2026-03-21T06:53:33.100441+0000 mgr.x (mgr.14152) 256 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:34.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:33 vm04 bash[20194]: cephadm 2026-03-21T06:53:33.100441+0000 mgr.x (mgr.14152) 256 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:34.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:33 vm04 bash[20194]: cluster 2026-03-21T06:53:33.580325+0000 mgr.x (mgr.14152) 257 : cluster [DBG] pgmap v194: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:34.257 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:33 vm04 bash[20194]: cluster 2026-03-21T06:53:33.580325+0000 mgr.x (mgr.14152) 257 : cluster [DBG] pgmap v194: 1 pgs: 1 active+undersized+degraded; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail; 2/6 objects degraded (33.333%) 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.035941+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.035941+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.040388+0000 mon.a (mon.0) 608 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.040388+0000 mon.a (mon.0) 608 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.044701+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.044701+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.048470+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.048470+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.123379+0000 mon.a (mon.0) 611 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.123379+0000 mon.a (mon.0) 611 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.127989+0000 mon.a (mon.0) 612 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.297 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:35 vm07 bash[19945]: audit 2026-03-21T06:53:34.127989+0000 mon.a (mon.0) 612 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.035941+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.035941+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.040388+0000 mon.a (mon.0) 608 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.040388+0000 mon.a (mon.0) 608 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.044701+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.044701+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.048470+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.048470+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.123379+0000 mon.a (mon.0) 611 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.123379+0000 mon.a (mon.0) 611 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.127989+0000 mon.a (mon.0) 612 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.391 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:35 vm02 bash[17657]: audit 2026-03-21T06:53:34.127989+0000 mon.a (mon.0) 612 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.035941+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.035941+0000 mon.a (mon.0) 607 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.040388+0000 mon.a (mon.0) 608 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.040388+0000 mon.a (mon.0) 608 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.044701+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.044701+0000 mon.a (mon.0) 609 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.048470+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.501 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.048470+0000 mon.a (mon.0) 610 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.502 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.123379+0000 mon.a (mon.0) 611 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.502 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.123379+0000 mon.a (mon.0) 611 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.502 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.127989+0000 mon.a (mon.0) 612 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.502 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:35 vm04 bash[20194]: audit 2026-03-21T06:53:34.127989+0000 mon.a (mon.0) 612 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:35.904 INFO:teuthology.orchestra.run.vm07.stdout:Created osd(s) 6 on host 'vm07' 2026-03-21T06:53:35.957 DEBUG:teuthology.orchestra.run.vm07:osd.6> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.6.service 2026-03-21T06:53:35.958 INFO:tasks.cephadm:Deploying osd.7 on vm07 with /dev/vg_nvme/lv_2... 2026-03-21T06:53:35.958 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 ceph-volume -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- lvm zap /dev/vg_nvme/lv_2 2026-03-21T06:53:36.151 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:36 vm07 bash[28634]: debug 2026-03-21T06:53:36.006+0000 7f5c6dd918c0 -1 Falling back to public interface 2026-03-21T06:53:36.230 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:53:36.534 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:36 vm07 bash[28634]: debug 2026-03-21T06:53:36.274+0000 7f5c6dd918c0 -1 osd.6 0 log_to_monitors true 2026-03-21T06:53:36.750 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:53:36.766 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch daemon add osd vm07:vg_nvme/lv_2 --skip-validation 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.532496+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.532496+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.538533+0000 mon.a (mon.0) 614 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.538533+0000 mon.a (mon.0) 614 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.556994+0000 mon.a (mon.0) 615 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.556994+0000 mon.a (mon.0) 615 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.562688+0000 mon.a (mon.0) 616 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.562688+0000 mon.a (mon.0) 616 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.567732+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.567732+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.572733+0000 mon.a (mon.0) 618 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.572733+0000 mon.a (mon.0) 618 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: cluster 2026-03-21T06:53:35.580604+0000 mgr.x (mgr.14152) 258 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: cluster 2026-03-21T06:53:35.580604+0000 mgr.x (mgr.14152) 258 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.899892+0000 mon.a (mon.0) 619 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.899892+0000 mon.a (mon.0) 619 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.904023+0000 mon.a (mon.0) 620 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:35.904023+0000 mon.a (mon.0) 620 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:36.281088+0000 mon.c (mon.1) 16 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:36.281088+0000 mon.c (mon.1) 16 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:36.281241+0000 mon.a (mon.0) 621 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:36.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:36 vm02 bash[17657]: audit 2026-03-21T06:53:36.281241+0000 mon.a (mon.0) 621 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.532496+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.532496+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.538533+0000 mon.a (mon.0) 614 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.538533+0000 mon.a (mon.0) 614 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.556994+0000 mon.a (mon.0) 615 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.556994+0000 mon.a (mon.0) 615 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.562688+0000 mon.a (mon.0) 616 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.562688+0000 mon.a (mon.0) 616 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.567732+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.567732+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.572733+0000 mon.a (mon.0) 618 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.572733+0000 mon.a (mon.0) 618 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: cluster 2026-03-21T06:53:35.580604+0000 mgr.x (mgr.14152) 258 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: cluster 2026-03-21T06:53:35.580604+0000 mgr.x (mgr.14152) 258 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.899892+0000 mon.a (mon.0) 619 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.899892+0000 mon.a (mon.0) 619 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.904023+0000 mon.a (mon.0) 620 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:35.904023+0000 mon.a (mon.0) 620 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:36.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:36.281088+0000 mon.c (mon.1) 16 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:36.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:36.281088+0000 mon.c (mon.1) 16 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:36.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:36.281241+0000 mon.a (mon.0) 621 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:36.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:36 vm07 bash[19945]: audit 2026-03-21T06:53:36.281241+0000 mon.a (mon.0) 621 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:37.004 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:53:37.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.532496+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.532496+0000 mon.a (mon.0) 613 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.538533+0000 mon.a (mon.0) 614 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.538533+0000 mon.a (mon.0) 614 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.556994+0000 mon.a (mon.0) 615 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.556994+0000 mon.a (mon.0) 615 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.562688+0000 mon.a (mon.0) 616 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.562688+0000 mon.a (mon.0) 616 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.567732+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.567732+0000 mon.a (mon.0) 617 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.572733+0000 mon.a (mon.0) 618 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.572733+0000 mon.a (mon.0) 618 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: cluster 2026-03-21T06:53:35.580604+0000 mgr.x (mgr.14152) 258 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: cluster 2026-03-21T06:53:35.580604+0000 mgr.x (mgr.14152) 258 : cluster [DBG] pgmap v195: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.899892+0000 mon.a (mon.0) 619 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.899892+0000 mon.a (mon.0) 619 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.904023+0000 mon.a (mon.0) 620 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:35.904023+0000 mon.a (mon.0) 620 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:36.281088+0000 mon.c (mon.1) 16 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:36.281088+0000 mon.c (mon.1) 16 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:36.281241+0000 mon.a (mon.0) 621 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:36 vm04 bash[20194]: audit 2026-03-21T06:53:36.281241+0000 mon.a (mon.0) 621 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: cluster 2026-03-21T06:53:36.571635+0000 mon.a (mon.0) 622 : cluster [INF] Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded) 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: cluster 2026-03-21T06:53:36.571635+0000 mon.a (mon.0) 622 : cluster [INF] Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded) 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: cluster 2026-03-21T06:53:36.571658+0000 mon.a (mon.0) 623 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: cluster 2026-03-21T06:53:36.571658+0000 mon.a (mon.0) 623 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:36.906770+0000 mon.a (mon.0) 624 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:36.906770+0000 mon.a (mon.0) 624 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: cluster 2026-03-21T06:53:36.908616+0000 mon.a (mon.0) 625 : cluster [DBG] osdmap e43: 7 total, 6 up, 7 in 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: cluster 2026-03-21T06:53:36.908616+0000 mon.a (mon.0) 625 : cluster [DBG] osdmap e43: 7 total, 6 up, 7 in 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:36.908794+0000 mon.a (mon.0) 626 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:36.908794+0000 mon.a (mon.0) 626 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:36.910568+0000 mon.c (mon.1) 17 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:36.910568+0000 mon.c (mon.1) 17 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:36.910727+0000 mon.a (mon.0) 627 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:36.910727+0000 mon.a (mon.0) 627 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.330722+0000 mon.a (mon.0) 628 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.330722+0000 mon.a (mon.0) 628 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.334676+0000 mon.a (mon.0) 629 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.334676+0000 mon.a (mon.0) 629 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.354313+0000 mon.a (mon.0) 630 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.354313+0000 mon.a (mon.0) 630 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.358207+0000 mon.a (mon.0) 631 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.358207+0000 mon.a (mon.0) 631 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.359046+0000 mon.a (mon.0) 632 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.359046+0000 mon.a (mon.0) 632 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.360581+0000 mon.a (mon.0) 633 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.360581+0000 mon.a (mon.0) 633 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.361198+0000 mon.a (mon.0) 634 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:37.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:37 vm02 bash[17657]: audit 2026-03-21T06:53:37.361198+0000 mon.a (mon.0) 634 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:37.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: cluster 2026-03-21T06:53:36.571635+0000 mon.a (mon.0) 622 : cluster [INF] Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded) 2026-03-21T06:53:37.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: cluster 2026-03-21T06:53:36.571635+0000 mon.a (mon.0) 622 : cluster [INF] Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded) 2026-03-21T06:53:37.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: cluster 2026-03-21T06:53:36.571658+0000 mon.a (mon.0) 623 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:37.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: cluster 2026-03-21T06:53:36.571658+0000 mon.a (mon.0) 623 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:36.906770+0000 mon.a (mon.0) 624 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:36.906770+0000 mon.a (mon.0) 624 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: cluster 2026-03-21T06:53:36.908616+0000 mon.a (mon.0) 625 : cluster [DBG] osdmap e43: 7 total, 6 up, 7 in 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: cluster 2026-03-21T06:53:36.908616+0000 mon.a (mon.0) 625 : cluster [DBG] osdmap e43: 7 total, 6 up, 7 in 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:36.908794+0000 mon.a (mon.0) 626 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:36.908794+0000 mon.a (mon.0) 626 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:36.910568+0000 mon.c (mon.1) 17 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:36.910568+0000 mon.c (mon.1) 17 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:36.910727+0000 mon.a (mon.0) 627 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:36.910727+0000 mon.a (mon.0) 627 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.330722+0000 mon.a (mon.0) 628 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.330722+0000 mon.a (mon.0) 628 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.334676+0000 mon.a (mon.0) 629 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.334676+0000 mon.a (mon.0) 629 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.354313+0000 mon.a (mon.0) 630 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.354313+0000 mon.a (mon.0) 630 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.358207+0000 mon.a (mon.0) 631 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.358207+0000 mon.a (mon.0) 631 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.359046+0000 mon.a (mon.0) 632 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.359046+0000 mon.a (mon.0) 632 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.360581+0000 mon.a (mon.0) 633 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.360581+0000 mon.a (mon.0) 633 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.361198+0000 mon.a (mon.0) 634 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:37.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:37 vm07 bash[19945]: audit 2026-03-21T06:53:37.361198+0000 mon.a (mon.0) 634 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:38.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: cluster 2026-03-21T06:53:36.571635+0000 mon.a (mon.0) 622 : cluster [INF] Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded) 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: cluster 2026-03-21T06:53:36.571635+0000 mon.a (mon.0) 622 : cluster [INF] Health check cleared: PG_DEGRADED (was: Degraded data redundancy: 2/6 objects degraded (33.333%), 1 pg degraded) 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: cluster 2026-03-21T06:53:36.571658+0000 mon.a (mon.0) 623 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: cluster 2026-03-21T06:53:36.571658+0000 mon.a (mon.0) 623 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:36.906770+0000 mon.a (mon.0) 624 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:36.906770+0000 mon.a (mon.0) 624 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: cluster 2026-03-21T06:53:36.908616+0000 mon.a (mon.0) 625 : cluster [DBG] osdmap e43: 7 total, 6 up, 7 in 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: cluster 2026-03-21T06:53:36.908616+0000 mon.a (mon.0) 625 : cluster [DBG] osdmap e43: 7 total, 6 up, 7 in 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:36.908794+0000 mon.a (mon.0) 626 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:36.908794+0000 mon.a (mon.0) 626 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:36.910568+0000 mon.c (mon.1) 17 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:36.910568+0000 mon.c (mon.1) 17 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:36.910727+0000 mon.a (mon.0) 627 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:36.910727+0000 mon.a (mon.0) 627 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.330722+0000 mon.a (mon.0) 628 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.330722+0000 mon.a (mon.0) 628 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.334676+0000 mon.a (mon.0) 629 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.334676+0000 mon.a (mon.0) 629 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.354313+0000 mon.a (mon.0) 630 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.354313+0000 mon.a (mon.0) 630 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.358207+0000 mon.a (mon.0) 631 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.358207+0000 mon.a (mon.0) 631 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.359046+0000 mon.a (mon.0) 632 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.359046+0000 mon.a (mon.0) 632 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.360581+0000 mon.a (mon.0) 633 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.360581+0000 mon.a (mon.0) 633 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.361198+0000 mon.a (mon.0) 634 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:37 vm04 bash[20194]: audit 2026-03-21T06:53:37.361198+0000 mon.a (mon.0) 634 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:38.401 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:37 vm07 bash[28634]: debug 2026-03-21T06:53:37.938+0000 7f5c69d11640 -1 osd.6 0 waiting for initial osdmap 2026-03-21T06:53:38.401 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:37 vm07 bash[28634]: debug 2026-03-21T06:53:37.942+0000 7f5c64af8640 -1 osd.6 44 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:37.342504+0000 mgr.x (mgr.14152) 259 : audit [DBG] from='client.24301 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:37.342504+0000 mgr.x (mgr.14152) 259 : audit [DBG] from='client.24301 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cephadm 2026-03-21T06:53:37.343243+0000 mgr.x (mgr.14152) 260 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cephadm 2026-03-21T06:53:37.343243+0000 mgr.x (mgr.14152) 260 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cephadm 2026-03-21T06:53:37.343302+0000 mgr.x (mgr.14152) 261 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_2'] 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cephadm 2026-03-21T06:53:37.343302+0000 mgr.x (mgr.14152) 261 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_2'] 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cephadm 2026-03-21T06:53:37.354650+0000 mgr.x (mgr.14152) 262 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cephadm 2026-03-21T06:53:37.354650+0000 mgr.x (mgr.14152) 262 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cephadm 2026-03-21T06:53:37.354783+0000 mgr.x (mgr.14152) 263 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cephadm 2026-03-21T06:53:37.354783+0000 mgr.x (mgr.14152) 263 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cluster 2026-03-21T06:53:37.580851+0000 mgr.x (mgr.14152) 264 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cluster 2026-03-21T06:53:37.580851+0000 mgr.x (mgr.14152) 264 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:37.910600+0000 mon.a (mon.0) 635 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:37.910600+0000 mon.a (mon.0) 635 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cluster 2026-03-21T06:53:37.912994+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e44: 7 total, 6 up, 7 in 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cluster 2026-03-21T06:53:37.912994+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e44: 7 total, 6 up, 7 in 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:37.913462+0000 mon.a (mon.0) 637 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:37.913462+0000 mon.a (mon.0) 637 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:37.918053+0000 mon.a (mon.0) 638 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:37.918053+0000 mon.a (mon.0) 638 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:38.244575+0000 mon.c (mon.1) 18 : audit [INF] from='client.? 192.168.123.107:0/171458073' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:38.244575+0000 mon.c (mon.1) 18 : audit [INF] from='client.? 192.168.123.107:0/171458073' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:38.244950+0000 mon.a (mon.0) 639 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:38.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:38.244950+0000 mon.a (mon.0) 639 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:38.892 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:38.248541+0000 mon.a (mon.0) 640 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"}]': finished 2026-03-21T06:53:38.892 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:38.248541+0000 mon.a (mon.0) 640 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"}]': finished 2026-03-21T06:53:38.892 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cluster 2026-03-21T06:53:38.250981+0000 mon.a (mon.0) 641 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504] boot 2026-03-21T06:53:38.892 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cluster 2026-03-21T06:53:38.250981+0000 mon.a (mon.0) 641 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504] boot 2026-03-21T06:53:38.892 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cluster 2026-03-21T06:53:38.251015+0000 mon.a (mon.0) 642 : cluster [DBG] osdmap e45: 8 total, 7 up, 8 in 2026-03-21T06:53:38.892 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: cluster 2026-03-21T06:53:38.251015+0000 mon.a (mon.0) 642 : cluster [DBG] osdmap e45: 8 total, 7 up, 8 in 2026-03-21T06:53:38.892 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:38.251407+0000 mon.a (mon.0) 643 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.892 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:38.251407+0000 mon.a (mon.0) 643 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.892 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:38.251526+0000 mon.a (mon.0) 644 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:38.892 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:38 vm02 bash[17657]: audit 2026-03-21T06:53:38.251526+0000 mon.a (mon.0) 644 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:38.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:37.342504+0000 mgr.x (mgr.14152) 259 : audit [DBG] from='client.24301 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:37.342504+0000 mgr.x (mgr.14152) 259 : audit [DBG] from='client.24301 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cephadm 2026-03-21T06:53:37.343243+0000 mgr.x (mgr.14152) 260 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cephadm 2026-03-21T06:53:37.343243+0000 mgr.x (mgr.14152) 260 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cephadm 2026-03-21T06:53:37.343302+0000 mgr.x (mgr.14152) 261 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_2'] 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cephadm 2026-03-21T06:53:37.343302+0000 mgr.x (mgr.14152) 261 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_2'] 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cephadm 2026-03-21T06:53:37.354650+0000 mgr.x (mgr.14152) 262 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cephadm 2026-03-21T06:53:37.354650+0000 mgr.x (mgr.14152) 262 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cephadm 2026-03-21T06:53:37.354783+0000 mgr.x (mgr.14152) 263 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cephadm 2026-03-21T06:53:37.354783+0000 mgr.x (mgr.14152) 263 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cluster 2026-03-21T06:53:37.580851+0000 mgr.x (mgr.14152) 264 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cluster 2026-03-21T06:53:37.580851+0000 mgr.x (mgr.14152) 264 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:37.910600+0000 mon.a (mon.0) 635 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:37.910600+0000 mon.a (mon.0) 635 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cluster 2026-03-21T06:53:37.912994+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e44: 7 total, 6 up, 7 in 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cluster 2026-03-21T06:53:37.912994+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e44: 7 total, 6 up, 7 in 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:37.913462+0000 mon.a (mon.0) 637 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:37.913462+0000 mon.a (mon.0) 637 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:37.918053+0000 mon.a (mon.0) 638 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:37.918053+0000 mon.a (mon.0) 638 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:38.244575+0000 mon.c (mon.1) 18 : audit [INF] from='client.? 192.168.123.107:0/171458073' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:38.244575+0000 mon.c (mon.1) 18 : audit [INF] from='client.? 192.168.123.107:0/171458073' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:38.244950+0000 mon.a (mon.0) 639 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:38.244950+0000 mon.a (mon.0) 639 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:38.248541+0000 mon.a (mon.0) 640 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"}]': finished 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:38.248541+0000 mon.a (mon.0) 640 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"}]': finished 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cluster 2026-03-21T06:53:38.250981+0000 mon.a (mon.0) 641 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504] boot 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cluster 2026-03-21T06:53:38.250981+0000 mon.a (mon.0) 641 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504] boot 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cluster 2026-03-21T06:53:38.251015+0000 mon.a (mon.0) 642 : cluster [DBG] osdmap e45: 8 total, 7 up, 8 in 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: cluster 2026-03-21T06:53:38.251015+0000 mon.a (mon.0) 642 : cluster [DBG] osdmap e45: 8 total, 7 up, 8 in 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:38.251407+0000 mon.a (mon.0) 643 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:38.251407+0000 mon.a (mon.0) 643 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:38.251526+0000 mon.a (mon.0) 644 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:38.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:38 vm07 bash[19945]: audit 2026-03-21T06:53:38.251526+0000 mon.a (mon.0) 644 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:37.342504+0000 mgr.x (mgr.14152) 259 : audit [DBG] from='client.24301 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:37.342504+0000 mgr.x (mgr.14152) 259 : audit [DBG] from='client.24301 -' entity='client.admin' cmd=[{"prefix": "orch daemon add osd", "svc_arg": "vm07:vg_nvme/lv_2", "skip_validation": true, "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cephadm 2026-03-21T06:53:37.343243+0000 mgr.x (mgr.14152) 260 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cephadm 2026-03-21T06:53:37.343243+0000 mgr.x (mgr.14152) 260 : cephadm [INF] osd.default does not exist. Creating it now. 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cephadm 2026-03-21T06:53:37.343302+0000 mgr.x (mgr.14152) 261 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_2'] 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cephadm 2026-03-21T06:53:37.343302+0000 mgr.x (mgr.14152) 261 : cephadm [INF] Creating OSDs with service ID: default on vm07:['vg_nvme/lv_2'] 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cephadm 2026-03-21T06:53:37.354650+0000 mgr.x (mgr.14152) 262 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cephadm 2026-03-21T06:53:37.354650+0000 mgr.x (mgr.14152) 262 : cephadm [INF] Marking host: vm07 for OSDSpec preview refresh. 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cephadm 2026-03-21T06:53:37.354783+0000 mgr.x (mgr.14152) 263 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cephadm 2026-03-21T06:53:37.354783+0000 mgr.x (mgr.14152) 263 : cephadm [INF] Saving service osd.default spec with placement vm07 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cluster 2026-03-21T06:53:37.580851+0000 mgr.x (mgr.14152) 264 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cluster 2026-03-21T06:53:37.580851+0000 mgr.x (mgr.14152) 264 : cluster [DBG] pgmap v197: 1 pgs: 1 active+clean; 449 KiB data, 162 MiB used, 120 GiB / 120 GiB avail 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:37.910600+0000 mon.a (mon.0) 635 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:37.910600+0000 mon.a (mon.0) 635 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cluster 2026-03-21T06:53:37.912994+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e44: 7 total, 6 up, 7 in 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cluster 2026-03-21T06:53:37.912994+0000 mon.a (mon.0) 636 : cluster [DBG] osdmap e44: 7 total, 6 up, 7 in 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:37.913462+0000 mon.a (mon.0) 637 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:37.913462+0000 mon.a (mon.0) 637 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:37.918053+0000 mon.a (mon.0) 638 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:37.918053+0000 mon.a (mon.0) 638 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:38.244575+0000 mon.c (mon.1) 18 : audit [INF] from='client.? 192.168.123.107:0/171458073' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:38.244575+0000 mon.c (mon.1) 18 : audit [INF] from='client.? 192.168.123.107:0/171458073' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:38.244950+0000 mon.a (mon.0) 639 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:38.244950+0000 mon.a (mon.0) 639 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd={"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:38.248541+0000 mon.a (mon.0) 640 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"}]': finished 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:38.248541+0000 mon.a (mon.0) 640 : audit [INF] from='client.? ' entity='client.bootstrap-osd' cmd='[{"prefix": "osd new", "uuid": "8e806dd2-e80e-40ab-8996-ff252b7491d9"}]': finished 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cluster 2026-03-21T06:53:38.250981+0000 mon.a (mon.0) 641 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504] boot 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cluster 2026-03-21T06:53:38.250981+0000 mon.a (mon.0) 641 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2331027504,v1:192.168.123.107:6809/2331027504] boot 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cluster 2026-03-21T06:53:38.251015+0000 mon.a (mon.0) 642 : cluster [DBG] osdmap e45: 8 total, 7 up, 8 in 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: cluster 2026-03-21T06:53:38.251015+0000 mon.a (mon.0) 642 : cluster [DBG] osdmap e45: 8 total, 7 up, 8 in 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:38.251407+0000 mon.a (mon.0) 643 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:38.251407+0000 mon.a (mon.0) 643 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:38.251526+0000 mon.a (mon.0) 644 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:38 vm04 bash[20194]: audit 2026-03-21T06:53:38.251526+0000 mon.a (mon.0) 644 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:39.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:39 vm02 bash[17657]: cluster 2026-03-21T06:53:37.274123+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:39.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:39 vm02 bash[17657]: cluster 2026-03-21T06:53:37.274123+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:39.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:39 vm02 bash[17657]: cluster 2026-03-21T06:53:37.274134+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:39.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:39 vm02 bash[17657]: cluster 2026-03-21T06:53:37.274134+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:39.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:39 vm02 bash[17657]: audit 2026-03-21T06:53:38.609238+0000 mon.c (mon.1) 19 : audit [DBG] from='client.? 192.168.123.107:0/1731066656' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:39.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:39 vm02 bash[17657]: audit 2026-03-21T06:53:38.609238+0000 mon.c (mon.1) 19 : audit [DBG] from='client.? 192.168.123.107:0/1731066656' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:39.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:39 vm02 bash[17657]: cluster 2026-03-21T06:53:39.254286+0000 mon.a (mon.0) 645 : cluster [DBG] osdmap e46: 8 total, 7 up, 8 in 2026-03-21T06:53:39.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:39 vm02 bash[17657]: cluster 2026-03-21T06:53:39.254286+0000 mon.a (mon.0) 645 : cluster [DBG] osdmap e46: 8 total, 7 up, 8 in 2026-03-21T06:53:39.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:39 vm02 bash[17657]: audit 2026-03-21T06:53:39.254414+0000 mon.a (mon.0) 646 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:39.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:39 vm02 bash[17657]: audit 2026-03-21T06:53:39.254414+0000 mon.a (mon.0) 646 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:39 vm07 bash[19945]: cluster 2026-03-21T06:53:37.274123+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:39 vm07 bash[19945]: cluster 2026-03-21T06:53:37.274123+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:39 vm07 bash[19945]: cluster 2026-03-21T06:53:37.274134+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:39 vm07 bash[19945]: cluster 2026-03-21T06:53:37.274134+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:39 vm07 bash[19945]: audit 2026-03-21T06:53:38.609238+0000 mon.c (mon.1) 19 : audit [DBG] from='client.? 192.168.123.107:0/1731066656' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:39 vm07 bash[19945]: audit 2026-03-21T06:53:38.609238+0000 mon.c (mon.1) 19 : audit [DBG] from='client.? 192.168.123.107:0/1731066656' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:39 vm07 bash[19945]: cluster 2026-03-21T06:53:39.254286+0000 mon.a (mon.0) 645 : cluster [DBG] osdmap e46: 8 total, 7 up, 8 in 2026-03-21T06:53:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:39 vm07 bash[19945]: cluster 2026-03-21T06:53:39.254286+0000 mon.a (mon.0) 645 : cluster [DBG] osdmap e46: 8 total, 7 up, 8 in 2026-03-21T06:53:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:39 vm07 bash[19945]: audit 2026-03-21T06:53:39.254414+0000 mon.a (mon.0) 646 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:39.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:39 vm07 bash[19945]: audit 2026-03-21T06:53:39.254414+0000 mon.a (mon.0) 646 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:39 vm04 bash[20194]: cluster 2026-03-21T06:53:37.274123+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:39 vm04 bash[20194]: cluster 2026-03-21T06:53:37.274123+0000 osd.6 (osd.6) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:39 vm04 bash[20194]: cluster 2026-03-21T06:53:37.274134+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:39 vm04 bash[20194]: cluster 2026-03-21T06:53:37.274134+0000 osd.6 (osd.6) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:39 vm04 bash[20194]: audit 2026-03-21T06:53:38.609238+0000 mon.c (mon.1) 19 : audit [DBG] from='client.? 192.168.123.107:0/1731066656' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:39 vm04 bash[20194]: audit 2026-03-21T06:53:38.609238+0000 mon.c (mon.1) 19 : audit [DBG] from='client.? 192.168.123.107:0/1731066656' entity='client.bootstrap-osd' cmd={"prefix": "mon getmap"} : dispatch 2026-03-21T06:53:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:39 vm04 bash[20194]: cluster 2026-03-21T06:53:39.254286+0000 mon.a (mon.0) 645 : cluster [DBG] osdmap e46: 8 total, 7 up, 8 in 2026-03-21T06:53:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:39 vm04 bash[20194]: cluster 2026-03-21T06:53:39.254286+0000 mon.a (mon.0) 645 : cluster [DBG] osdmap e46: 8 total, 7 up, 8 in 2026-03-21T06:53:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:39 vm04 bash[20194]: audit 2026-03-21T06:53:39.254414+0000 mon.a (mon.0) 646 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:39 vm04 bash[20194]: audit 2026-03-21T06:53:39.254414+0000 mon.a (mon.0) 646 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:40.605 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: cluster 2026-03-21T06:53:39.581167+0000 mgr.x (mgr.14152) 265 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:40.605 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: cluster 2026-03-21T06:53:39.581167+0000 mgr.x (mgr.14152) 265 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:40.605 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.376565+0000 mon.a (mon.0) 647 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.605 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.376565+0000 mon.a (mon.0) 647 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.605 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.382239+0000 mon.a (mon.0) 648 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.605 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.382239+0000 mon.a (mon.0) 648 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.605 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.387876+0000 mon.a (mon.0) 649 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.605 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.387876+0000 mon.a (mon.0) 649 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.605 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.392735+0000 mon.a (mon.0) 650 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.605 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.392735+0000 mon.a (mon.0) 650 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.605 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.393555+0000 mon.a (mon.0) 651 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.393555+0000 mon.a (mon.0) 651 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.397507+0000 mon.a (mon.0) 652 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.397507+0000 mon.a (mon.0) 652 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.399019+0000 mon.a (mon.0) 653 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.399019+0000 mon.a (mon.0) 653 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.399578+0000 mon.a (mon.0) 654 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.399578+0000 mon.a (mon.0) 654 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.403737+0000 mon.a (mon.0) 655 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.403737+0000 mon.a (mon.0) 655 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.404741+0000 mon.a (mon.0) 656 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.404741+0000 mon.a (mon.0) 656 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.406114+0000 mon.a (mon.0) 657 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.406114+0000 mon.a (mon.0) 657 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.406580+0000 mon.a (mon.0) 658 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:40.606 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:40 vm07 bash[19945]: audit 2026-03-21T06:53:40.406580+0000 mon.a (mon.0) 658 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: cluster 2026-03-21T06:53:39.581167+0000 mgr.x (mgr.14152) 265 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: cluster 2026-03-21T06:53:39.581167+0000 mgr.x (mgr.14152) 265 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.376565+0000 mon.a (mon.0) 647 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.376565+0000 mon.a (mon.0) 647 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.382239+0000 mon.a (mon.0) 648 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.382239+0000 mon.a (mon.0) 648 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.387876+0000 mon.a (mon.0) 649 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.387876+0000 mon.a (mon.0) 649 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.392735+0000 mon.a (mon.0) 650 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.392735+0000 mon.a (mon.0) 650 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.393555+0000 mon.a (mon.0) 651 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.393555+0000 mon.a (mon.0) 651 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.397507+0000 mon.a (mon.0) 652 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.397507+0000 mon.a (mon.0) 652 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.399019+0000 mon.a (mon.0) 653 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.399019+0000 mon.a (mon.0) 653 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.399578+0000 mon.a (mon.0) 654 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.399578+0000 mon.a (mon.0) 654 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.403737+0000 mon.a (mon.0) 655 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.403737+0000 mon.a (mon.0) 655 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.404741+0000 mon.a (mon.0) 656 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.404741+0000 mon.a (mon.0) 656 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.406114+0000 mon.a (mon.0) 657 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.406114+0000 mon.a (mon.0) 657 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.406580+0000 mon.a (mon.0) 658 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:40.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:40 vm02 bash[17657]: audit 2026-03-21T06:53:40.406580+0000 mon.a (mon.0) 658 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: cluster 2026-03-21T06:53:39.581167+0000 mgr.x (mgr.14152) 265 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: cluster 2026-03-21T06:53:39.581167+0000 mgr.x (mgr.14152) 265 : cluster [DBG] pgmap v201: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.376565+0000 mon.a (mon.0) 647 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.376565+0000 mon.a (mon.0) 647 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.382239+0000 mon.a (mon.0) 648 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.382239+0000 mon.a (mon.0) 648 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.387876+0000 mon.a (mon.0) 649 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.387876+0000 mon.a (mon.0) 649 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.392735+0000 mon.a (mon.0) 650 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.392735+0000 mon.a (mon.0) 650 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.393555+0000 mon.a (mon.0) 651 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.393555+0000 mon.a (mon.0) 651 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.397507+0000 mon.a (mon.0) 652 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.397507+0000 mon.a (mon.0) 652 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.399019+0000 mon.a (mon.0) 653 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.399019+0000 mon.a (mon.0) 653 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.399578+0000 mon.a (mon.0) 654 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.399578+0000 mon.a (mon.0) 654 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.403737+0000 mon.a (mon.0) 655 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.403737+0000 mon.a (mon.0) 655 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.404741+0000 mon.a (mon.0) 656 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.404741+0000 mon.a (mon.0) 656 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.406114+0000 mon.a (mon.0) 657 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.406114+0000 mon.a (mon.0) 657 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.bootstrap-osd"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.406580+0000 mon.a (mon.0) 658 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:40 vm04 bash[20194]: audit 2026-03-21T06:53:40.406580+0000 mon.a (mon.0) 658 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:41.727 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:41.727 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:41.727 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:41 vm07 systemd[1]: Stopping Ceph osd.6 for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T06:53:41.727 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:41 vm07 bash[19945]: cephadm 2026-03-21T06:53:40.370006+0000 mgr.x (mgr.14152) 266 : cephadm [INF] Detected new or changed devices on vm07 2026-03-21T06:53:41.727 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:41 vm07 bash[19945]: cephadm 2026-03-21T06:53:40.370006+0000 mgr.x (mgr.14152) 266 : cephadm [INF] Detected new or changed devices on vm07 2026-03-21T06:53:41.727 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:41 vm07 bash[19945]: cephadm 2026-03-21T06:53:40.394075+0000 mgr.x (mgr.14152) 267 : cephadm [INF] Adjusting osd_memory_target on vm07 to 5934M 2026-03-21T06:53:41.727 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:41 vm07 bash[19945]: cephadm 2026-03-21T06:53:40.394075+0000 mgr.x (mgr.14152) 267 : cephadm [INF] Adjusting osd_memory_target on vm07 to 5934M 2026-03-21T06:53:41.727 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:41 vm07 bash[19945]: audit 2026-03-21T06:53:40.682436+0000 mon.a (mon.0) 659 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:41.727 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:41 vm07 bash[19945]: audit 2026-03-21T06:53:40.682436+0000 mon.a (mon.0) 659 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:41.727 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:41 vm07 bash[19945]: audit 2026-03-21T06:53:40.683259+0000 mon.a (mon.0) 660 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:41.727 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:41 vm07 bash[19945]: audit 2026-03-21T06:53:40.683259+0000 mon.a (mon.0) 660 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:41.727 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:41 vm07 bash[19945]: cephadm 2026-03-21T06:53:40.683687+0000 mgr.x (mgr.14152) 268 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:41.727 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:41 vm07 bash[19945]: cephadm 2026-03-21T06:53:40.683687+0000 mgr.x (mgr.14152) 268 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:41.727 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:41.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:41 vm02 bash[17657]: cephadm 2026-03-21T06:53:40.370006+0000 mgr.x (mgr.14152) 266 : cephadm [INF] Detected new or changed devices on vm07 2026-03-21T06:53:41.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:41 vm02 bash[17657]: cephadm 2026-03-21T06:53:40.370006+0000 mgr.x (mgr.14152) 266 : cephadm [INF] Detected new or changed devices on vm07 2026-03-21T06:53:41.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:41 vm02 bash[17657]: cephadm 2026-03-21T06:53:40.394075+0000 mgr.x (mgr.14152) 267 : cephadm [INF] Adjusting osd_memory_target on vm07 to 5934M 2026-03-21T06:53:41.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:41 vm02 bash[17657]: cephadm 2026-03-21T06:53:40.394075+0000 mgr.x (mgr.14152) 267 : cephadm [INF] Adjusting osd_memory_target on vm07 to 5934M 2026-03-21T06:53:41.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:41 vm02 bash[17657]: audit 2026-03-21T06:53:40.682436+0000 mon.a (mon.0) 659 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:41.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:41 vm02 bash[17657]: audit 2026-03-21T06:53:40.682436+0000 mon.a (mon.0) 659 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:41.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:41 vm02 bash[17657]: audit 2026-03-21T06:53:40.683259+0000 mon.a (mon.0) 660 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:41.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:41 vm02 bash[17657]: audit 2026-03-21T06:53:40.683259+0000 mon.a (mon.0) 660 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:41.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:41 vm02 bash[17657]: cephadm 2026-03-21T06:53:40.683687+0000 mgr.x (mgr.14152) 268 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:41.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:41 vm02 bash[17657]: cephadm 2026-03-21T06:53:40.683687+0000 mgr.x (mgr.14152) 268 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:42.007 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:41 vm04 bash[20194]: cephadm 2026-03-21T06:53:40.370006+0000 mgr.x (mgr.14152) 266 : cephadm [INF] Detected new or changed devices on vm07 2026-03-21T06:53:42.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:41 vm04 bash[20194]: cephadm 2026-03-21T06:53:40.370006+0000 mgr.x (mgr.14152) 266 : cephadm [INF] Detected new or changed devices on vm07 2026-03-21T06:53:42.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:41 vm04 bash[20194]: cephadm 2026-03-21T06:53:40.394075+0000 mgr.x (mgr.14152) 267 : cephadm [INF] Adjusting osd_memory_target on vm07 to 5934M 2026-03-21T06:53:42.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:41 vm04 bash[20194]: cephadm 2026-03-21T06:53:40.394075+0000 mgr.x (mgr.14152) 267 : cephadm [INF] Adjusting osd_memory_target on vm07 to 5934M 2026-03-21T06:53:42.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:41 vm04 bash[20194]: audit 2026-03-21T06:53:40.682436+0000 mon.a (mon.0) 659 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:42.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:41 vm04 bash[20194]: audit 2026-03-21T06:53:40.682436+0000 mon.a (mon.0) 659 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.6"} : dispatch 2026-03-21T06:53:42.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:41 vm04 bash[20194]: audit 2026-03-21T06:53:40.683259+0000 mon.a (mon.0) 660 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:42.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:41 vm04 bash[20194]: audit 2026-03-21T06:53:40.683259+0000 mon.a (mon.0) 660 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:42.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:41 vm04 bash[20194]: cephadm 2026-03-21T06:53:40.683687+0000 mgr.x (mgr.14152) 268 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:42.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:41 vm04 bash[20194]: cephadm 2026-03-21T06:53:40.683687+0000 mgr.x (mgr.14152) 268 : cephadm [INF] Deploying daemon osd.6 on vm07 2026-03-21T06:53:42.151 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:41 vm07 bash[28634]: debug 2026-03-21T06:53:41.722+0000 7f5c6ad26640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.6 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-21T06:53:42.151 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:41 vm07 bash[28634]: debug 2026-03-21T06:53:41.722+0000 7f5c6ad26640 -1 osd.6 46 *** Got signal Terminated *** 2026-03-21T06:53:42.151 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:41 vm07 bash[28634]: debug 2026-03-21T06:53:41.722+0000 7f5c6ad26640 -1 osd.6 46 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-21T06:53:42.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:42 vm02 bash[17657]: cluster 2026-03-21T06:53:41.581464+0000 mgr.x (mgr.14152) 269 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:42.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:42 vm02 bash[17657]: cluster 2026-03-21T06:53:41.581464+0000 mgr.x (mgr.14152) 269 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:42.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:42 vm02 bash[17657]: cluster 2026-03-21T06:53:41.727567+0000 mon.a (mon.0) 661 : cluster [INF] osd.6 marked itself down and dead 2026-03-21T06:53:42.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:42 vm02 bash[17657]: cluster 2026-03-21T06:53:41.727567+0000 mon.a (mon.0) 661 : cluster [INF] osd.6 marked itself down and dead 2026-03-21T06:53:42.901 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:42 vm07 bash[34198]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-6 2026-03-21T06:53:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:42 vm07 bash[19945]: cluster 2026-03-21T06:53:41.581464+0000 mgr.x (mgr.14152) 269 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:42 vm07 bash[19945]: cluster 2026-03-21T06:53:41.581464+0000 mgr.x (mgr.14152) 269 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:42 vm07 bash[19945]: cluster 2026-03-21T06:53:41.727567+0000 mon.a (mon.0) 661 : cluster [INF] osd.6 marked itself down and dead 2026-03-21T06:53:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:42 vm07 bash[19945]: cluster 2026-03-21T06:53:41.727567+0000 mon.a (mon.0) 661 : cluster [INF] osd.6 marked itself down and dead 2026-03-21T06:53:43.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:42 vm04 bash[20194]: cluster 2026-03-21T06:53:41.581464+0000 mgr.x (mgr.14152) 269 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:43.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:42 vm04 bash[20194]: cluster 2026-03-21T06:53:41.581464+0000 mgr.x (mgr.14152) 269 : cluster [DBG] pgmap v202: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:43.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:42 vm04 bash[20194]: cluster 2026-03-21T06:53:41.727567+0000 mon.a (mon.0) 661 : cluster [INF] osd.6 marked itself down and dead 2026-03-21T06:53:43.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:42 vm04 bash[20194]: cluster 2026-03-21T06:53:41.727567+0000 mon.a (mon.0) 661 : cluster [INF] osd.6 marked itself down and dead 2026-03-21T06:53:43.161 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:43 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:43.161 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:42 vm07 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.6.service: Deactivated successfully. 2026-03-21T06:53:43.161 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:42 vm07 systemd[1]: Stopped Ceph osd.6 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T06:53:43.161 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:43 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:43.161 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:43 vm07 systemd[1]: Started Ceph osd.6 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T06:53:43.161 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:43.470 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:43 vm07 bash[34417]: Running command: /usr/bin/ceph-authtool --gen-print-key 2026-03-21T06:53:43.470 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:43 vm07 bash[34417]: Running command: /usr/bin/ceph-authtool --gen-print-key 2026-03-21T06:53:43.728 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: cluster 2026-03-21T06:53:42.555285+0000 mon.a (mon.0) 662 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: cluster 2026-03-21T06:53:42.555285+0000 mon.a (mon.0) 662 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: cluster 2026-03-21T06:53:42.568395+0000 mon.a (mon.0) 663 : cluster [DBG] osdmap e47: 8 total, 6 up, 8 in 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: cluster 2026-03-21T06:53:42.568395+0000 mon.a (mon.0) 663 : cluster [DBG] osdmap e47: 8 total, 6 up, 8 in 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:42.568559+0000 mon.a (mon.0) 664 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:42.568559+0000 mon.a (mon.0) 664 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.120338+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.120338+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.124935+0000 mon.a (mon.0) 666 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.124935+0000 mon.a (mon.0) 666 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.125505+0000 mon.a (mon.0) 667 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.125505+0000 mon.a (mon.0) 667 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.126047+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.126047+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.534547+0000 mon.a (mon.0) 669 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.534547+0000 mon.a (mon.0) 669 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.535085+0000 mon.a (mon.0) 670 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:43.729 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:43 vm07 bash[19945]: audit 2026-03-21T06:53:43.535085+0000 mon.a (mon.0) 670 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: cluster 2026-03-21T06:53:42.555285+0000 mon.a (mon.0) 662 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: cluster 2026-03-21T06:53:42.555285+0000 mon.a (mon.0) 662 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: cluster 2026-03-21T06:53:42.568395+0000 mon.a (mon.0) 663 : cluster [DBG] osdmap e47: 8 total, 6 up, 8 in 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: cluster 2026-03-21T06:53:42.568395+0000 mon.a (mon.0) 663 : cluster [DBG] osdmap e47: 8 total, 6 up, 8 in 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:42.568559+0000 mon.a (mon.0) 664 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:42.568559+0000 mon.a (mon.0) 664 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.120338+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.120338+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.124935+0000 mon.a (mon.0) 666 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.124935+0000 mon.a (mon.0) 666 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.125505+0000 mon.a (mon.0) 667 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.125505+0000 mon.a (mon.0) 667 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.126047+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.126047+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.534547+0000 mon.a (mon.0) 669 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.534547+0000 mon.a (mon.0) 669 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.535085+0000 mon.a (mon.0) 670 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:43.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:43 vm02 bash[17657]: audit 2026-03-21T06:53:43.535085+0000 mon.a (mon.0) 670 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: cluster 2026-03-21T06:53:42.555285+0000 mon.a (mon.0) 662 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: cluster 2026-03-21T06:53:42.555285+0000 mon.a (mon.0) 662 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: cluster 2026-03-21T06:53:42.568395+0000 mon.a (mon.0) 663 : cluster [DBG] osdmap e47: 8 total, 6 up, 8 in 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: cluster 2026-03-21T06:53:42.568395+0000 mon.a (mon.0) 663 : cluster [DBG] osdmap e47: 8 total, 6 up, 8 in 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:42.568559+0000 mon.a (mon.0) 664 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:42.568559+0000 mon.a (mon.0) 664 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.120338+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.120338+0000 mon.a (mon.0) 665 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.124935+0000 mon.a (mon.0) 666 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.124935+0000 mon.a (mon.0) 666 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.125505+0000 mon.a (mon.0) 667 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.125505+0000 mon.a (mon.0) 667 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.126047+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.126047+0000 mon.a (mon.0) 668 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.534547+0000 mon.a (mon.0) 669 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.534547+0000 mon.a (mon.0) 669 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "osd.7"} : dispatch 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.535085+0000 mon.a (mon.0) 670 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:44.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:43 vm04 bash[20194]: audit 2026-03-21T06:53:43.535085+0000 mon.a (mon.0) 670 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:53:44.314 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:44 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:44.314 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:44.314 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:44.314 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 bash[34417]: --> Failed to activate via raw: did not find any matching OSD to activate 2026-03-21T06:53:44.314 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 bash[34417]: Running command: /usr/bin/ceph-authtool --gen-print-key 2026-03-21T06:53:44.314 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 bash[34417]: Running command: /usr/bin/ceph-authtool --gen-print-key 2026-03-21T06:53:44.567 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:44 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:44.567 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:44.567 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:44.567 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 bash[34417]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-6 2026-03-21T06:53:44.567 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 bash[34417]: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/vg_nvme/lv_3 --path /var/lib/ceph/osd/ceph-6 --no-mon-config 2026-03-21T06:53:44.567 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 bash[34417]: Running command: /usr/bin/ln -snf /dev/vg_nvme/lv_3 /var/lib/ceph/osd/ceph-6/block 2026-03-21T06:53:44.567 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 bash[34417]: Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-6/block 2026-03-21T06:53:44.567 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 bash[34417]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-2 2026-03-21T06:53:44.567 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 bash[34417]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-6 2026-03-21T06:53:44.567 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:44 vm07 bash[34417]: --> ceph-volume lvm activate successful for osd ID: 6 2026-03-21T06:53:44.824 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 bash[19945]: cephadm 2026-03-21T06:53:43.126429+0000 mgr.x (mgr.14152) 270 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:44.824 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 bash[19945]: cephadm 2026-03-21T06:53:43.126429+0000 mgr.x (mgr.14152) 270 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:44.824 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 bash[19945]: cephadm 2026-03-21T06:53:43.535476+0000 mgr.x (mgr.14152) 271 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:44.824 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 bash[19945]: cephadm 2026-03-21T06:53:43.535476+0000 mgr.x (mgr.14152) 271 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:44.824 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 bash[19945]: cluster 2026-03-21T06:53:43.581700+0000 mgr.x (mgr.14152) 272 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:44.824 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 bash[19945]: cluster 2026-03-21T06:53:43.581700+0000 mgr.x (mgr.14152) 272 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:44.824 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 bash[19945]: audit 2026-03-21T06:53:44.428717+0000 mon.a (mon.0) 671 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:44.824 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 bash[19945]: audit 2026-03-21T06:53:44.428717+0000 mon.a (mon.0) 671 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:44.824 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 bash[19945]: audit 2026-03-21T06:53:44.433798+0000 mon.a (mon.0) 672 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:44.824 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:44 vm07 bash[19945]: audit 2026-03-21T06:53:44.433798+0000 mon.a (mon.0) 672 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:44.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:44 vm02 bash[17657]: cephadm 2026-03-21T06:53:43.126429+0000 mgr.x (mgr.14152) 270 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:44.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:44 vm02 bash[17657]: cephadm 2026-03-21T06:53:43.126429+0000 mgr.x (mgr.14152) 270 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:44.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:44 vm02 bash[17657]: cephadm 2026-03-21T06:53:43.535476+0000 mgr.x (mgr.14152) 271 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:44.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:44 vm02 bash[17657]: cephadm 2026-03-21T06:53:43.535476+0000 mgr.x (mgr.14152) 271 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:44.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:44 vm02 bash[17657]: cluster 2026-03-21T06:53:43.581700+0000 mgr.x (mgr.14152) 272 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:44.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:44 vm02 bash[17657]: cluster 2026-03-21T06:53:43.581700+0000 mgr.x (mgr.14152) 272 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:44.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:44 vm02 bash[17657]: audit 2026-03-21T06:53:44.428717+0000 mon.a (mon.0) 671 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:44.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:44 vm02 bash[17657]: audit 2026-03-21T06:53:44.428717+0000 mon.a (mon.0) 671 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:44.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:44 vm02 bash[17657]: audit 2026-03-21T06:53:44.433798+0000 mon.a (mon.0) 672 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:44.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:44 vm02 bash[17657]: audit 2026-03-21T06:53:44.433798+0000 mon.a (mon.0) 672 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:44 vm04 bash[20194]: cephadm 2026-03-21T06:53:43.126429+0000 mgr.x (mgr.14152) 270 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:44 vm04 bash[20194]: cephadm 2026-03-21T06:53:43.126429+0000 mgr.x (mgr.14152) 270 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:44 vm04 bash[20194]: cephadm 2026-03-21T06:53:43.535476+0000 mgr.x (mgr.14152) 271 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:44 vm04 bash[20194]: cephadm 2026-03-21T06:53:43.535476+0000 mgr.x (mgr.14152) 271 : cephadm [INF] Deploying daemon osd.7 on vm07 2026-03-21T06:53:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:44 vm04 bash[20194]: cluster 2026-03-21T06:53:43.581700+0000 mgr.x (mgr.14152) 272 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:44 vm04 bash[20194]: cluster 2026-03-21T06:53:43.581700+0000 mgr.x (mgr.14152) 272 : cluster [DBG] pgmap v204: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:44 vm04 bash[20194]: audit 2026-03-21T06:53:44.428717+0000 mon.a (mon.0) 671 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:44 vm04 bash[20194]: audit 2026-03-21T06:53:44.428717+0000 mon.a (mon.0) 671 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:44 vm04 bash[20194]: audit 2026-03-21T06:53:44.433798+0000 mon.a (mon.0) 672 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:44 vm04 bash[20194]: audit 2026-03-21T06:53:44.433798+0000 mon.a (mon.0) 672 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:53:45.151 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:45 vm07 bash[35243]: debug 2026-03-21T06:53:45.050+0000 7fefac5c48c0 -1 Falling back to public interface 2026-03-21T06:53:45.546 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:45 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:45.546 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:53:45 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:45.546 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:45 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:53:45.546 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:45 vm07 bash[35243]: debug 2026-03-21T06:53:45.342+0000 7fefac5c48c0 -1 osd.6 46 log_to_monitors true 2026-03-21T06:53:45.831 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:45 vm07 bash[19945]: audit 2026-03-21T06:53:45.348993+0000 mon.c (mon.1) 20 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:45.832 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:45 vm07 bash[19945]: audit 2026-03-21T06:53:45.348993+0000 mon.c (mon.1) 20 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:45.832 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:45 vm07 bash[19945]: audit 2026-03-21T06:53:45.349423+0000 mon.a (mon.0) 673 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:45.832 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:45 vm07 bash[19945]: audit 2026-03-21T06:53:45.349423+0000 mon.a (mon.0) 673 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:45.832 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:53:45 vm07 bash[35243]: debug 2026-03-21T06:53:45.622+0000 7fefa332b640 -1 osd.6 46 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-21T06:53:45.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:45 vm02 bash[17657]: audit 2026-03-21T06:53:45.348993+0000 mon.c (mon.1) 20 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:45.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:45 vm02 bash[17657]: audit 2026-03-21T06:53:45.348993+0000 mon.c (mon.1) 20 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:45.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:45 vm02 bash[17657]: audit 2026-03-21T06:53:45.349423+0000 mon.a (mon.0) 673 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:45.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:45 vm02 bash[17657]: audit 2026-03-21T06:53:45.349423+0000 mon.a (mon.0) 673 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:46.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:45 vm04 bash[20194]: audit 2026-03-21T06:53:45.348993+0000 mon.c (mon.1) 20 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:46.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:45 vm04 bash[20194]: audit 2026-03-21T06:53:45.348993+0000 mon.c (mon.1) 20 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:46.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:45 vm04 bash[20194]: audit 2026-03-21T06:53:45.349423+0000 mon.a (mon.0) 673 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:46.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:45 vm04 bash[20194]: audit 2026-03-21T06:53:45.349423+0000 mon.a (mon.0) 673 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]} : dispatch 2026-03-21T06:53:46.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: cluster 2026-03-21T06:53:45.581998+0000 mgr.x (mgr.14152) 273 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:46.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: cluster 2026-03-21T06:53:45.581998+0000 mgr.x (mgr.14152) 273 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:46.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: audit 2026-03-21T06:53:45.590177+0000 mon.a (mon.0) 674 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:46.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: audit 2026-03-21T06:53:45.590177+0000 mon.a (mon.0) 674 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:46.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: cluster 2026-03-21T06:53:45.593406+0000 mon.a (mon.0) 675 : cluster [DBG] osdmap e48: 8 total, 6 up, 8 in 2026-03-21T06:53:46.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: cluster 2026-03-21T06:53:45.593406+0000 mon.a (mon.0) 675 : cluster [DBG] osdmap e48: 8 total, 6 up, 8 in 2026-03-21T06:53:46.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: audit 2026-03-21T06:53:45.593561+0000 mon.a (mon.0) 676 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:46.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: audit 2026-03-21T06:53:45.593561+0000 mon.a (mon.0) 676 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:46.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: audit 2026-03-21T06:53:45.597677+0000 mon.c (mon.1) 21 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:46.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: audit 2026-03-21T06:53:45.597677+0000 mon.c (mon.1) 21 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:46.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: audit 2026-03-21T06:53:45.597941+0000 mon.a (mon.0) 677 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:46.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:46 vm02 bash[17657]: audit 2026-03-21T06:53:45.597941+0000 mon.a (mon.0) 677 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: cluster 2026-03-21T06:53:45.581998+0000 mgr.x (mgr.14152) 273 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: cluster 2026-03-21T06:53:45.581998+0000 mgr.x (mgr.14152) 273 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: audit 2026-03-21T06:53:45.590177+0000 mon.a (mon.0) 674 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: audit 2026-03-21T06:53:45.590177+0000 mon.a (mon.0) 674 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: cluster 2026-03-21T06:53:45.593406+0000 mon.a (mon.0) 675 : cluster [DBG] osdmap e48: 8 total, 6 up, 8 in 2026-03-21T06:53:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: cluster 2026-03-21T06:53:45.593406+0000 mon.a (mon.0) 675 : cluster [DBG] osdmap e48: 8 total, 6 up, 8 in 2026-03-21T06:53:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: audit 2026-03-21T06:53:45.593561+0000 mon.a (mon.0) 676 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:46.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: audit 2026-03-21T06:53:45.593561+0000 mon.a (mon.0) 676 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:46.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: audit 2026-03-21T06:53:45.597677+0000 mon.c (mon.1) 21 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:46.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: audit 2026-03-21T06:53:45.597677+0000 mon.c (mon.1) 21 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:46.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: audit 2026-03-21T06:53:45.597941+0000 mon.a (mon.0) 677 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:46.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:46 vm07 bash[19945]: audit 2026-03-21T06:53:45.597941+0000 mon.a (mon.0) 677 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: cluster 2026-03-21T06:53:45.581998+0000 mgr.x (mgr.14152) 273 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: cluster 2026-03-21T06:53:45.581998+0000 mgr.x (mgr.14152) 273 : cluster [DBG] pgmap v205: 1 pgs: 1 active+clean; 449 KiB data, 189 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: audit 2026-03-21T06:53:45.590177+0000 mon.a (mon.0) 674 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: audit 2026-03-21T06:53:45.590177+0000 mon.a (mon.0) 674 : audit [INF] from='osd.6 ' entity='osd.6' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["6"]}]': finished 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: cluster 2026-03-21T06:53:45.593406+0000 mon.a (mon.0) 675 : cluster [DBG] osdmap e48: 8 total, 6 up, 8 in 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: cluster 2026-03-21T06:53:45.593406+0000 mon.a (mon.0) 675 : cluster [DBG] osdmap e48: 8 total, 6 up, 8 in 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: audit 2026-03-21T06:53:45.593561+0000 mon.a (mon.0) 676 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: audit 2026-03-21T06:53:45.593561+0000 mon.a (mon.0) 676 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: audit 2026-03-21T06:53:45.597677+0000 mon.c (mon.1) 21 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: audit 2026-03-21T06:53:45.597677+0000 mon.c (mon.1) 21 : audit [INF] from='osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328]' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: audit 2026-03-21T06:53:45.597941+0000 mon.a (mon.0) 677 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:46 vm04 bash[20194]: audit 2026-03-21T06:53:45.597941+0000 mon.a (mon.0) 677 : audit [INF] from='osd.6 ' entity='osd.6' cmd={"prefix": "osd crush create-or-move", "id": 6, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: cluster 2026-03-21T06:53:46.590513+0000 mon.a (mon.0) 678 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: cluster 2026-03-21T06:53:46.590513+0000 mon.a (mon.0) 678 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: cluster 2026-03-21T06:53:46.590535+0000 mon.a (mon.0) 679 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: cluster 2026-03-21T06:53:46.590535+0000 mon.a (mon.0) 679 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: cluster 2026-03-21T06:53:46.595771+0000 mon.a (mon.0) 680 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328] boot 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: cluster 2026-03-21T06:53:46.595771+0000 mon.a (mon.0) 680 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328] boot 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: cluster 2026-03-21T06:53:46.595799+0000 mon.a (mon.0) 681 : cluster [DBG] osdmap e49: 8 total, 7 up, 8 in 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: cluster 2026-03-21T06:53:46.595799+0000 mon.a (mon.0) 681 : cluster [DBG] osdmap e49: 8 total, 7 up, 8 in 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: audit 2026-03-21T06:53:46.596487+0000 mon.a (mon.0) 682 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: audit 2026-03-21T06:53:46.596487+0000 mon.a (mon.0) 682 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: audit 2026-03-21T06:53:46.596636+0000 mon.a (mon.0) 683 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: audit 2026-03-21T06:53:46.596636+0000 mon.a (mon.0) 683 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: audit 2026-03-21T06:53:46.613545+0000 mon.c (mon.1) 22 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: audit 2026-03-21T06:53:46.613545+0000 mon.c (mon.1) 22 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: audit 2026-03-21T06:53:46.613859+0000 mon.a (mon.0) 684 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:47.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:47 vm02 bash[17657]: audit 2026-03-21T06:53:46.613859+0000 mon.a (mon.0) 684 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: cluster 2026-03-21T06:53:46.590513+0000 mon.a (mon.0) 678 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: cluster 2026-03-21T06:53:46.590513+0000 mon.a (mon.0) 678 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: cluster 2026-03-21T06:53:46.590535+0000 mon.a (mon.0) 679 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: cluster 2026-03-21T06:53:46.590535+0000 mon.a (mon.0) 679 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: cluster 2026-03-21T06:53:46.595771+0000 mon.a (mon.0) 680 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328] boot 2026-03-21T06:53:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: cluster 2026-03-21T06:53:46.595771+0000 mon.a (mon.0) 680 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328] boot 2026-03-21T06:53:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: cluster 2026-03-21T06:53:46.595799+0000 mon.a (mon.0) 681 : cluster [DBG] osdmap e49: 8 total, 7 up, 8 in 2026-03-21T06:53:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: cluster 2026-03-21T06:53:46.595799+0000 mon.a (mon.0) 681 : cluster [DBG] osdmap e49: 8 total, 7 up, 8 in 2026-03-21T06:53:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: audit 2026-03-21T06:53:46.596487+0000 mon.a (mon.0) 682 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: audit 2026-03-21T06:53:46.596487+0000 mon.a (mon.0) 682 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: audit 2026-03-21T06:53:46.596636+0000 mon.a (mon.0) 683 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: audit 2026-03-21T06:53:46.596636+0000 mon.a (mon.0) 683 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: audit 2026-03-21T06:53:46.613545+0000 mon.c (mon.1) 22 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: audit 2026-03-21T06:53:46.613545+0000 mon.c (mon.1) 22 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: audit 2026-03-21T06:53:46.613859+0000 mon.a (mon.0) 684 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:47 vm07 bash[19945]: audit 2026-03-21T06:53:46.613859+0000 mon.a (mon.0) 684 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: cluster 2026-03-21T06:53:46.590513+0000 mon.a (mon.0) 678 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: cluster 2026-03-21T06:53:46.590513+0000 mon.a (mon.0) 678 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: cluster 2026-03-21T06:53:46.590535+0000 mon.a (mon.0) 679 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: cluster 2026-03-21T06:53:46.590535+0000 mon.a (mon.0) 679 : cluster [INF] Cluster is now healthy 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: cluster 2026-03-21T06:53:46.595771+0000 mon.a (mon.0) 680 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328] boot 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: cluster 2026-03-21T06:53:46.595771+0000 mon.a (mon.0) 680 : cluster [INF] osd.6 [v2:192.168.123.107:6808/2090727328,v1:192.168.123.107:6809/2090727328] boot 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: cluster 2026-03-21T06:53:46.595799+0000 mon.a (mon.0) 681 : cluster [DBG] osdmap e49: 8 total, 7 up, 8 in 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: cluster 2026-03-21T06:53:46.595799+0000 mon.a (mon.0) 681 : cluster [DBG] osdmap e49: 8 total, 7 up, 8 in 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: audit 2026-03-21T06:53:46.596487+0000 mon.a (mon.0) 682 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: audit 2026-03-21T06:53:46.596487+0000 mon.a (mon.0) 682 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 6} : dispatch 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: audit 2026-03-21T06:53:46.596636+0000 mon.a (mon.0) 683 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: audit 2026-03-21T06:53:46.596636+0000 mon.a (mon.0) 683 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: audit 2026-03-21T06:53:46.613545+0000 mon.c (mon.1) 22 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: audit 2026-03-21T06:53:46.613545+0000 mon.c (mon.1) 22 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: audit 2026-03-21T06:53:46.613859+0000 mon.a (mon.0) 684 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:47 vm04 bash[20194]: audit 2026-03-21T06:53:46.613859+0000 mon.a (mon.0) 684 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: cluster 2026-03-21T06:53:47.582320+0000 mgr.x (mgr.14152) 274 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: cluster 2026-03-21T06:53:47.582320+0000 mgr.x (mgr.14152) 274 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: audit 2026-03-21T06:53:47.610146+0000 mon.a (mon.0) 685 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: audit 2026-03-21T06:53:47.610146+0000 mon.a (mon.0) 685 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: cluster 2026-03-21T06:53:47.611712+0000 mon.a (mon.0) 686 : cluster [DBG] osdmap e50: 8 total, 7 up, 8 in 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: cluster 2026-03-21T06:53:47.611712+0000 mon.a (mon.0) 686 : cluster [DBG] osdmap e50: 8 total, 7 up, 8 in 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: audit 2026-03-21T06:53:47.611858+0000 mon.a (mon.0) 687 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: audit 2026-03-21T06:53:47.611858+0000 mon.a (mon.0) 687 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: audit 2026-03-21T06:53:47.613520+0000 mon.c (mon.1) 23 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: audit 2026-03-21T06:53:47.613520+0000 mon.c (mon.1) 23 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: audit 2026-03-21T06:53:47.613779+0000 mon.a (mon.0) 688 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:48.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:48 vm02 bash[17657]: audit 2026-03-21T06:53:47.613779+0000 mon.a (mon.0) 688 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: cluster 2026-03-21T06:53:47.582320+0000 mgr.x (mgr.14152) 274 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: cluster 2026-03-21T06:53:47.582320+0000 mgr.x (mgr.14152) 274 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: audit 2026-03-21T06:53:47.610146+0000 mon.a (mon.0) 685 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: audit 2026-03-21T06:53:47.610146+0000 mon.a (mon.0) 685 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: cluster 2026-03-21T06:53:47.611712+0000 mon.a (mon.0) 686 : cluster [DBG] osdmap e50: 8 total, 7 up, 8 in 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: cluster 2026-03-21T06:53:47.611712+0000 mon.a (mon.0) 686 : cluster [DBG] osdmap e50: 8 total, 7 up, 8 in 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: audit 2026-03-21T06:53:47.611858+0000 mon.a (mon.0) 687 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: audit 2026-03-21T06:53:47.611858+0000 mon.a (mon.0) 687 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: audit 2026-03-21T06:53:47.613520+0000 mon.c (mon.1) 23 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: audit 2026-03-21T06:53:47.613520+0000 mon.c (mon.1) 23 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: audit 2026-03-21T06:53:47.613779+0000 mon.a (mon.0) 688 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:48 vm07 bash[19945]: audit 2026-03-21T06:53:47.613779+0000 mon.a (mon.0) 688 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: cluster 2026-03-21T06:53:47.582320+0000 mgr.x (mgr.14152) 274 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: cluster 2026-03-21T06:53:47.582320+0000 mgr.x (mgr.14152) 274 : cluster [DBG] pgmap v208: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: audit 2026-03-21T06:53:47.610146+0000 mon.a (mon.0) 685 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: audit 2026-03-21T06:53:47.610146+0000 mon.a (mon.0) 685 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: cluster 2026-03-21T06:53:47.611712+0000 mon.a (mon.0) 686 : cluster [DBG] osdmap e50: 8 total, 7 up, 8 in 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: cluster 2026-03-21T06:53:47.611712+0000 mon.a (mon.0) 686 : cluster [DBG] osdmap e50: 8 total, 7 up, 8 in 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: audit 2026-03-21T06:53:47.611858+0000 mon.a (mon.0) 687 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: audit 2026-03-21T06:53:47.611858+0000 mon.a (mon.0) 687 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: audit 2026-03-21T06:53:47.613520+0000 mon.c (mon.1) 23 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: audit 2026-03-21T06:53:47.613520+0000 mon.c (mon.1) 23 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: audit 2026-03-21T06:53:47.613779+0000 mon.a (mon.0) 688 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:48 vm04 bash[20194]: audit 2026-03-21T06:53:47.613779+0000 mon.a (mon.0) 688 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:53:49.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: cluster 2026-03-21T06:53:47.648084+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:49.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: cluster 2026-03-21T06:53:47.648084+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:49.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: cluster 2026-03-21T06:53:47.648091+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:49.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: cluster 2026-03-21T06:53:47.648091+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:49.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: audit 2026-03-21T06:53:48.613621+0000 mon.a (mon.0) 689 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:49.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: audit 2026-03-21T06:53:48.613621+0000 mon.a (mon.0) 689 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:49.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: cluster 2026-03-21T06:53:48.615957+0000 mon.a (mon.0) 690 : cluster [DBG] osdmap e51: 8 total, 7 up, 8 in 2026-03-21T06:53:49.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: cluster 2026-03-21T06:53:48.615957+0000 mon.a (mon.0) 690 : cluster [DBG] osdmap e51: 8 total, 7 up, 8 in 2026-03-21T06:53:49.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: audit 2026-03-21T06:53:48.616663+0000 mon.a (mon.0) 691 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:49.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: audit 2026-03-21T06:53:48.616663+0000 mon.a (mon.0) 691 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:49.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: audit 2026-03-21T06:53:48.621522+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:49.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:49 vm02 bash[17657]: audit 2026-03-21T06:53:48.621522+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: cluster 2026-03-21T06:53:47.648084+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: cluster 2026-03-21T06:53:47.648084+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: cluster 2026-03-21T06:53:47.648091+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: cluster 2026-03-21T06:53:47.648091+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: audit 2026-03-21T06:53:48.613621+0000 mon.a (mon.0) 689 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: audit 2026-03-21T06:53:48.613621+0000 mon.a (mon.0) 689 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: cluster 2026-03-21T06:53:48.615957+0000 mon.a (mon.0) 690 : cluster [DBG] osdmap e51: 8 total, 7 up, 8 in 2026-03-21T06:53:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: cluster 2026-03-21T06:53:48.615957+0000 mon.a (mon.0) 690 : cluster [DBG] osdmap e51: 8 total, 7 up, 8 in 2026-03-21T06:53:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: audit 2026-03-21T06:53:48.616663+0000 mon.a (mon.0) 691 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:49.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: audit 2026-03-21T06:53:48.616663+0000 mon.a (mon.0) 691 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:49.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: audit 2026-03-21T06:53:48.621522+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:49.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:49 vm07 bash[19945]: audit 2026-03-21T06:53:48.621522+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: cluster 2026-03-21T06:53:47.648084+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: cluster 2026-03-21T06:53:47.648084+0000 osd.7 (osd.7) 1 : cluster [DBG] purged_snaps scrub starts 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: cluster 2026-03-21T06:53:47.648091+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: cluster 2026-03-21T06:53:47.648091+0000 osd.7 (osd.7) 2 : cluster [DBG] purged_snaps scrub ok 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: audit 2026-03-21T06:53:48.613621+0000 mon.a (mon.0) 689 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: audit 2026-03-21T06:53:48.613621+0000 mon.a (mon.0) 689 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]}]': finished 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: cluster 2026-03-21T06:53:48.615957+0000 mon.a (mon.0) 690 : cluster [DBG] osdmap e51: 8 total, 7 up, 8 in 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: cluster 2026-03-21T06:53:48.615957+0000 mon.a (mon.0) 690 : cluster [DBG] osdmap e51: 8 total, 7 up, 8 in 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: audit 2026-03-21T06:53:48.616663+0000 mon.a (mon.0) 691 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: audit 2026-03-21T06:53:48.616663+0000 mon.a (mon.0) 691 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: audit 2026-03-21T06:53:48.621522+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:49 vm04 bash[20194]: audit 2026-03-21T06:53:48.621522+0000 mon.a (mon.0) 692 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:50 vm02 bash[17657]: cluster 2026-03-21T06:53:49.582619+0000 mgr.x (mgr.14152) 275 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:50.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:50 vm02 bash[17657]: cluster 2026-03-21T06:53:49.582619+0000 mgr.x (mgr.14152) 275 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:50.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:50 vm02 bash[17657]: audit 2026-03-21T06:53:49.619705+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:50 vm02 bash[17657]: audit 2026-03-21T06:53:49.619705+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:50 vm02 bash[17657]: cluster 2026-03-21T06:53:49.629583+0000 mon.a (mon.0) 694 : cluster [INF] osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005] boot 2026-03-21T06:53:50.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:50 vm02 bash[17657]: cluster 2026-03-21T06:53:49.629583+0000 mon.a (mon.0) 694 : cluster [INF] osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005] boot 2026-03-21T06:53:50.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:50 vm02 bash[17657]: cluster 2026-03-21T06:53:49.629647+0000 mon.a (mon.0) 695 : cluster [DBG] osdmap e52: 8 total, 8 up, 8 in 2026-03-21T06:53:50.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:50 vm02 bash[17657]: cluster 2026-03-21T06:53:49.629647+0000 mon.a (mon.0) 695 : cluster [DBG] osdmap e52: 8 total, 8 up, 8 in 2026-03-21T06:53:50.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:50 vm02 bash[17657]: audit 2026-03-21T06:53:49.629959+0000 mon.a (mon.0) 696 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:50 vm02 bash[17657]: audit 2026-03-21T06:53:49.629959+0000 mon.a (mon.0) 696 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:50 vm07 bash[19945]: cluster 2026-03-21T06:53:49.582619+0000 mgr.x (mgr.14152) 275 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:50 vm07 bash[19945]: cluster 2026-03-21T06:53:49.582619+0000 mgr.x (mgr.14152) 275 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:50 vm07 bash[19945]: audit 2026-03-21T06:53:49.619705+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:50 vm07 bash[19945]: audit 2026-03-21T06:53:49.619705+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:50 vm07 bash[19945]: cluster 2026-03-21T06:53:49.629583+0000 mon.a (mon.0) 694 : cluster [INF] osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005] boot 2026-03-21T06:53:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:50 vm07 bash[19945]: cluster 2026-03-21T06:53:49.629583+0000 mon.a (mon.0) 694 : cluster [INF] osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005] boot 2026-03-21T06:53:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:50 vm07 bash[19945]: cluster 2026-03-21T06:53:49.629647+0000 mon.a (mon.0) 695 : cluster [DBG] osdmap e52: 8 total, 8 up, 8 in 2026-03-21T06:53:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:50 vm07 bash[19945]: cluster 2026-03-21T06:53:49.629647+0000 mon.a (mon.0) 695 : cluster [DBG] osdmap e52: 8 total, 8 up, 8 in 2026-03-21T06:53:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:50 vm07 bash[19945]: audit 2026-03-21T06:53:49.629959+0000 mon.a (mon.0) 696 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:50.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:50 vm07 bash[19945]: audit 2026-03-21T06:53:49.629959+0000 mon.a (mon.0) 696 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:50 vm04 bash[20194]: cluster 2026-03-21T06:53:49.582619+0000 mgr.x (mgr.14152) 275 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:50 vm04 bash[20194]: cluster 2026-03-21T06:53:49.582619+0000 mgr.x (mgr.14152) 275 : cluster [DBG] pgmap v211: 1 pgs: 1 active+clean; 449 KiB data, 190 MiB used, 140 GiB / 140 GiB avail 2026-03-21T06:53:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:50 vm04 bash[20194]: audit 2026-03-21T06:53:49.619705+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:50 vm04 bash[20194]: audit 2026-03-21T06:53:49.619705+0000 mon.a (mon.0) 693 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:50 vm04 bash[20194]: cluster 2026-03-21T06:53:49.629583+0000 mon.a (mon.0) 694 : cluster [INF] osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005] boot 2026-03-21T06:53:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:50 vm04 bash[20194]: cluster 2026-03-21T06:53:49.629583+0000 mon.a (mon.0) 694 : cluster [INF] osd.7 [v2:192.168.123.107:6816/3920516005,v1:192.168.123.107:6817/3920516005] boot 2026-03-21T06:53:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:50 vm04 bash[20194]: cluster 2026-03-21T06:53:49.629647+0000 mon.a (mon.0) 695 : cluster [DBG] osdmap e52: 8 total, 8 up, 8 in 2026-03-21T06:53:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:50 vm04 bash[20194]: cluster 2026-03-21T06:53:49.629647+0000 mon.a (mon.0) 695 : cluster [DBG] osdmap e52: 8 total, 8 up, 8 in 2026-03-21T06:53:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:50 vm04 bash[20194]: audit 2026-03-21T06:53:49.629959+0000 mon.a (mon.0) 696 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:50 vm04 bash[20194]: audit 2026-03-21T06:53:49.629959+0000 mon.a (mon.0) 696 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:53:51.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:51 vm02 bash[17657]: cluster 2026-03-21T06:53:50.631833+0000 mon.a (mon.0) 697 : cluster [DBG] osdmap e53: 8 total, 8 up, 8 in 2026-03-21T06:53:51.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:51 vm02 bash[17657]: cluster 2026-03-21T06:53:50.631833+0000 mon.a (mon.0) 697 : cluster [DBG] osdmap e53: 8 total, 8 up, 8 in 2026-03-21T06:53:51.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:51 vm07 bash[19945]: cluster 2026-03-21T06:53:50.631833+0000 mon.a (mon.0) 697 : cluster [DBG] osdmap e53: 8 total, 8 up, 8 in 2026-03-21T06:53:51.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:51 vm07 bash[19945]: cluster 2026-03-21T06:53:50.631833+0000 mon.a (mon.0) 697 : cluster [DBG] osdmap e53: 8 total, 8 up, 8 in 2026-03-21T06:53:52.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:51 vm04 bash[20194]: cluster 2026-03-21T06:53:50.631833+0000 mon.a (mon.0) 697 : cluster [DBG] osdmap e53: 8 total, 8 up, 8 in 2026-03-21T06:53:52.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:51 vm04 bash[20194]: cluster 2026-03-21T06:53:50.631833+0000 mon.a (mon.0) 697 : cluster [DBG] osdmap e53: 8 total, 8 up, 8 in 2026-03-21T06:53:52.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:52 vm07 bash[19945]: cluster 2026-03-21T06:53:51.582928+0000 mgr.x (mgr.14152) 276 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:52.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:52 vm07 bash[19945]: cluster 2026-03-21T06:53:51.582928+0000 mgr.x (mgr.14152) 276 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:53.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:52 vm04 bash[20194]: cluster 2026-03-21T06:53:51.582928+0000 mgr.x (mgr.14152) 276 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:53.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:52 vm04 bash[20194]: cluster 2026-03-21T06:53:51.582928+0000 mgr.x (mgr.14152) 276 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:53.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:52 vm02 bash[17657]: cluster 2026-03-21T06:53:51.582928+0000 mgr.x (mgr.14152) 276 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:53.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:52 vm02 bash[17657]: cluster 2026-03-21T06:53:51.582928+0000 mgr.x (mgr.14152) 276 : cluster [DBG] pgmap v214: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:54.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:54 vm07 bash[19945]: cluster 2026-03-21T06:53:53.583418+0000 mgr.x (mgr.14152) 277 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:54.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:54 vm07 bash[19945]: cluster 2026-03-21T06:53:53.583418+0000 mgr.x (mgr.14152) 277 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:55.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:54 vm04 bash[20194]: cluster 2026-03-21T06:53:53.583418+0000 mgr.x (mgr.14152) 277 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:55.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:54 vm04 bash[20194]: cluster 2026-03-21T06:53:53.583418+0000 mgr.x (mgr.14152) 277 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:55.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:54 vm02 bash[17657]: cluster 2026-03-21T06:53:53.583418+0000 mgr.x (mgr.14152) 277 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:55.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:54 vm02 bash[17657]: cluster 2026-03-21T06:53:53.583418+0000 mgr.x (mgr.14152) 277 : cluster [DBG] pgmap v215: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:57.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:56 vm04 bash[20194]: cluster 2026-03-21T06:53:55.583747+0000 mgr.x (mgr.14152) 278 : cluster [DBG] pgmap v216: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:57.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:56 vm04 bash[20194]: cluster 2026-03-21T06:53:55.583747+0000 mgr.x (mgr.14152) 278 : cluster [DBG] pgmap v216: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:57.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:56 vm02 bash[17657]: cluster 2026-03-21T06:53:55.583747+0000 mgr.x (mgr.14152) 278 : cluster [DBG] pgmap v216: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:57.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:56 vm02 bash[17657]: cluster 2026-03-21T06:53:55.583747+0000 mgr.x (mgr.14152) 278 : cluster [DBG] pgmap v216: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:57.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:56 vm07 bash[19945]: cluster 2026-03-21T06:53:55.583747+0000 mgr.x (mgr.14152) 278 : cluster [DBG] pgmap v216: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:57.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:56 vm07 bash[19945]: cluster 2026-03-21T06:53:55.583747+0000 mgr.x (mgr.14152) 278 : cluster [DBG] pgmap v216: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:59.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:58 vm04 bash[20194]: cluster 2026-03-21T06:53:57.584083+0000 mgr.x (mgr.14152) 279 : cluster [DBG] pgmap v217: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:53:58 vm04 bash[20194]: cluster 2026-03-21T06:53:57.584083+0000 mgr.x (mgr.14152) 279 : cluster [DBG] pgmap v217: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:59.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:58 vm02 bash[17657]: cluster 2026-03-21T06:53:57.584083+0000 mgr.x (mgr.14152) 279 : cluster [DBG] pgmap v217: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:59.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:53:58 vm02 bash[17657]: cluster 2026-03-21T06:53:57.584083+0000 mgr.x (mgr.14152) 279 : cluster [DBG] pgmap v217: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:58 vm07 bash[19945]: cluster 2026-03-21T06:53:57.584083+0000 mgr.x (mgr.14152) 279 : cluster [DBG] pgmap v217: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:53:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:53:58 vm07 bash[19945]: cluster 2026-03-21T06:53:57.584083+0000 mgr.x (mgr.14152) 279 : cluster [DBG] pgmap v217: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:00 vm04 bash[20194]: cluster 2026-03-21T06:53:59.584410+0000 mgr.x (mgr.14152) 280 : cluster [DBG] pgmap v218: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:01.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:00 vm04 bash[20194]: cluster 2026-03-21T06:53:59.584410+0000 mgr.x (mgr.14152) 280 : cluster [DBG] pgmap v218: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:01.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:00 vm02 bash[17657]: cluster 2026-03-21T06:53:59.584410+0000 mgr.x (mgr.14152) 280 : cluster [DBG] pgmap v218: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:01.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:00 vm02 bash[17657]: cluster 2026-03-21T06:53:59.584410+0000 mgr.x (mgr.14152) 280 : cluster [DBG] pgmap v218: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:01.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:00 vm07 bash[19945]: cluster 2026-03-21T06:53:59.584410+0000 mgr.x (mgr.14152) 280 : cluster [DBG] pgmap v218: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:01.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:00 vm07 bash[19945]: cluster 2026-03-21T06:53:59.584410+0000 mgr.x (mgr.14152) 280 : cluster [DBG] pgmap v218: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:02 vm04 bash[20194]: cluster 2026-03-21T06:54:01.584709+0000 mgr.x (mgr.14152) 281 : cluster [DBG] pgmap v219: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:02 vm04 bash[20194]: cluster 2026-03-21T06:54:01.584709+0000 mgr.x (mgr.14152) 281 : cluster [DBG] pgmap v219: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:03.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:02 vm02 bash[17657]: cluster 2026-03-21T06:54:01.584709+0000 mgr.x (mgr.14152) 281 : cluster [DBG] pgmap v219: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:03.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:02 vm02 bash[17657]: cluster 2026-03-21T06:54:01.584709+0000 mgr.x (mgr.14152) 281 : cluster [DBG] pgmap v219: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:03.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:02 vm07 bash[19945]: cluster 2026-03-21T06:54:01.584709+0000 mgr.x (mgr.14152) 281 : cluster [DBG] pgmap v219: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:03.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:02 vm07 bash[19945]: cluster 2026-03-21T06:54:01.584709+0000 mgr.x (mgr.14152) 281 : cluster [DBG] pgmap v219: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:04 vm04 bash[20194]: cluster 2026-03-21T06:54:03.584991+0000 mgr.x (mgr.14152) 282 : cluster [DBG] pgmap v220: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:04 vm04 bash[20194]: cluster 2026-03-21T06:54:03.584991+0000 mgr.x (mgr.14152) 282 : cluster [DBG] pgmap v220: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:05.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:04 vm02 bash[17657]: cluster 2026-03-21T06:54:03.584991+0000 mgr.x (mgr.14152) 282 : cluster [DBG] pgmap v220: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:05.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:04 vm02 bash[17657]: cluster 2026-03-21T06:54:03.584991+0000 mgr.x (mgr.14152) 282 : cluster [DBG] pgmap v220: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:05.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:04 vm07 bash[19945]: cluster 2026-03-21T06:54:03.584991+0000 mgr.x (mgr.14152) 282 : cluster [DBG] pgmap v220: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:05.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:04 vm07 bash[19945]: cluster 2026-03-21T06:54:03.584991+0000 mgr.x (mgr.14152) 282 : cluster [DBG] pgmap v220: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:07.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:06 vm04 bash[20194]: cluster 2026-03-21T06:54:05.585295+0000 mgr.x (mgr.14152) 283 : cluster [DBG] pgmap v221: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:07.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:06 vm04 bash[20194]: cluster 2026-03-21T06:54:05.585295+0000 mgr.x (mgr.14152) 283 : cluster [DBG] pgmap v221: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:07.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:06 vm02 bash[17657]: cluster 2026-03-21T06:54:05.585295+0000 mgr.x (mgr.14152) 283 : cluster [DBG] pgmap v221: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:07.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:06 vm02 bash[17657]: cluster 2026-03-21T06:54:05.585295+0000 mgr.x (mgr.14152) 283 : cluster [DBG] pgmap v221: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:06 vm07 bash[19945]: cluster 2026-03-21T06:54:05.585295+0000 mgr.x (mgr.14152) 283 : cluster [DBG] pgmap v221: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:06 vm07 bash[19945]: cluster 2026-03-21T06:54:05.585295+0000 mgr.x (mgr.14152) 283 : cluster [DBG] pgmap v221: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:08 vm04 bash[20194]: cluster 2026-03-21T06:54:07.585672+0000 mgr.x (mgr.14152) 284 : cluster [DBG] pgmap v222: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:08 vm04 bash[20194]: cluster 2026-03-21T06:54:07.585672+0000 mgr.x (mgr.14152) 284 : cluster [DBG] pgmap v222: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:09.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:08 vm02 bash[17657]: cluster 2026-03-21T06:54:07.585672+0000 mgr.x (mgr.14152) 284 : cluster [DBG] pgmap v222: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:09.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:08 vm02 bash[17657]: cluster 2026-03-21T06:54:07.585672+0000 mgr.x (mgr.14152) 284 : cluster [DBG] pgmap v222: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:08 vm07 bash[19945]: cluster 2026-03-21T06:54:07.585672+0000 mgr.x (mgr.14152) 284 : cluster [DBG] pgmap v222: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:08 vm07 bash[19945]: cluster 2026-03-21T06:54:07.585672+0000 mgr.x (mgr.14152) 284 : cluster [DBG] pgmap v222: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:11.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:10 vm04 bash[20194]: cluster 2026-03-21T06:54:09.586137+0000 mgr.x (mgr.14152) 285 : cluster [DBG] pgmap v223: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:11.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:10 vm04 bash[20194]: cluster 2026-03-21T06:54:09.586137+0000 mgr.x (mgr.14152) 285 : cluster [DBG] pgmap v223: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:11.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:10 vm02 bash[17657]: cluster 2026-03-21T06:54:09.586137+0000 mgr.x (mgr.14152) 285 : cluster [DBG] pgmap v223: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:11.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:10 vm02 bash[17657]: cluster 2026-03-21T06:54:09.586137+0000 mgr.x (mgr.14152) 285 : cluster [DBG] pgmap v223: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:10 vm07 bash[19945]: cluster 2026-03-21T06:54:09.586137+0000 mgr.x (mgr.14152) 285 : cluster [DBG] pgmap v223: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:10 vm07 bash[19945]: cluster 2026-03-21T06:54:09.586137+0000 mgr.x (mgr.14152) 285 : cluster [DBG] pgmap v223: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:12.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:11 vm04 bash[20194]: cluster 2026-03-21T06:54:11.586491+0000 mgr.x (mgr.14152) 286 : cluster [DBG] pgmap v224: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:12.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:11 vm04 bash[20194]: cluster 2026-03-21T06:54:11.586491+0000 mgr.x (mgr.14152) 286 : cluster [DBG] pgmap v224: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:12.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:11 vm02 bash[17657]: cluster 2026-03-21T06:54:11.586491+0000 mgr.x (mgr.14152) 286 : cluster [DBG] pgmap v224: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:12.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:11 vm02 bash[17657]: cluster 2026-03-21T06:54:11.586491+0000 mgr.x (mgr.14152) 286 : cluster [DBG] pgmap v224: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:12.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:11 vm07 bash[19945]: cluster 2026-03-21T06:54:11.586491+0000 mgr.x (mgr.14152) 286 : cluster [DBG] pgmap v224: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:12.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:11 vm07 bash[19945]: cluster 2026-03-21T06:54:11.586491+0000 mgr.x (mgr.14152) 286 : cluster [DBG] pgmap v224: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:14.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:14 vm07 bash[19945]: cluster 2026-03-21T06:54:13.586844+0000 mgr.x (mgr.14152) 287 : cluster [DBG] pgmap v225: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:14.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:14 vm07 bash[19945]: cluster 2026-03-21T06:54:13.586844+0000 mgr.x (mgr.14152) 287 : cluster [DBG] pgmap v225: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:15.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:14 vm04 bash[20194]: cluster 2026-03-21T06:54:13.586844+0000 mgr.x (mgr.14152) 287 : cluster [DBG] pgmap v225: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:15.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:14 vm04 bash[20194]: cluster 2026-03-21T06:54:13.586844+0000 mgr.x (mgr.14152) 287 : cluster [DBG] pgmap v225: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:15.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:14 vm02 bash[17657]: cluster 2026-03-21T06:54:13.586844+0000 mgr.x (mgr.14152) 287 : cluster [DBG] pgmap v225: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:15.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:14 vm02 bash[17657]: cluster 2026-03-21T06:54:13.586844+0000 mgr.x (mgr.14152) 287 : cluster [DBG] pgmap v225: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:16.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:16 vm07 bash[19945]: cluster 2026-03-21T06:54:15.587208+0000 mgr.x (mgr.14152) 288 : cluster [DBG] pgmap v226: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:16.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:16 vm07 bash[19945]: cluster 2026-03-21T06:54:15.587208+0000 mgr.x (mgr.14152) 288 : cluster [DBG] pgmap v226: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:16 vm04 bash[20194]: cluster 2026-03-21T06:54:15.587208+0000 mgr.x (mgr.14152) 288 : cluster [DBG] pgmap v226: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:16 vm04 bash[20194]: cluster 2026-03-21T06:54:15.587208+0000 mgr.x (mgr.14152) 288 : cluster [DBG] pgmap v226: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:17.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:16 vm02 bash[17657]: cluster 2026-03-21T06:54:15.587208+0000 mgr.x (mgr.14152) 288 : cluster [DBG] pgmap v226: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:17.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:16 vm02 bash[17657]: cluster 2026-03-21T06:54:15.587208+0000 mgr.x (mgr.14152) 288 : cluster [DBG] pgmap v226: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:18 vm04 bash[20194]: cluster 2026-03-21T06:54:17.587633+0000 mgr.x (mgr.14152) 289 : cluster [DBG] pgmap v227: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:18 vm04 bash[20194]: cluster 2026-03-21T06:54:17.587633+0000 mgr.x (mgr.14152) 289 : cluster [DBG] pgmap v227: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:19.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:18 vm02 bash[17657]: cluster 2026-03-21T06:54:17.587633+0000 mgr.x (mgr.14152) 289 : cluster [DBG] pgmap v227: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:19.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:18 vm02 bash[17657]: cluster 2026-03-21T06:54:17.587633+0000 mgr.x (mgr.14152) 289 : cluster [DBG] pgmap v227: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:18 vm07 bash[19945]: cluster 2026-03-21T06:54:17.587633+0000 mgr.x (mgr.14152) 289 : cluster [DBG] pgmap v227: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:18 vm07 bash[19945]: cluster 2026-03-21T06:54:17.587633+0000 mgr.x (mgr.14152) 289 : cluster [DBG] pgmap v227: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:20 vm04 bash[20194]: cluster 2026-03-21T06:54:19.588036+0000 mgr.x (mgr.14152) 290 : cluster [DBG] pgmap v228: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:20 vm04 bash[20194]: cluster 2026-03-21T06:54:19.588036+0000 mgr.x (mgr.14152) 290 : cluster [DBG] pgmap v228: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:21.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:20 vm02 bash[17657]: cluster 2026-03-21T06:54:19.588036+0000 mgr.x (mgr.14152) 290 : cluster [DBG] pgmap v228: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:21.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:20 vm02 bash[17657]: cluster 2026-03-21T06:54:19.588036+0000 mgr.x (mgr.14152) 290 : cluster [DBG] pgmap v228: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:20 vm07 bash[19945]: cluster 2026-03-21T06:54:19.588036+0000 mgr.x (mgr.14152) 290 : cluster [DBG] pgmap v228: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:20 vm07 bash[19945]: cluster 2026-03-21T06:54:19.588036+0000 mgr.x (mgr.14152) 290 : cluster [DBG] pgmap v228: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:23.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:22 vm04 bash[20194]: cluster 2026-03-21T06:54:21.588344+0000 mgr.x (mgr.14152) 291 : cluster [DBG] pgmap v229: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:23.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:22 vm04 bash[20194]: cluster 2026-03-21T06:54:21.588344+0000 mgr.x (mgr.14152) 291 : cluster [DBG] pgmap v229: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:23.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:22 vm02 bash[17657]: cluster 2026-03-21T06:54:21.588344+0000 mgr.x (mgr.14152) 291 : cluster [DBG] pgmap v229: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:23.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:22 vm02 bash[17657]: cluster 2026-03-21T06:54:21.588344+0000 mgr.x (mgr.14152) 291 : cluster [DBG] pgmap v229: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:22 vm07 bash[19945]: cluster 2026-03-21T06:54:21.588344+0000 mgr.x (mgr.14152) 291 : cluster [DBG] pgmap v229: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:22 vm07 bash[19945]: cluster 2026-03-21T06:54:21.588344+0000 mgr.x (mgr.14152) 291 : cluster [DBG] pgmap v229: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:25.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:24 vm04 bash[20194]: cluster 2026-03-21T06:54:23.588680+0000 mgr.x (mgr.14152) 292 : cluster [DBG] pgmap v230: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:25.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:24 vm04 bash[20194]: cluster 2026-03-21T06:54:23.588680+0000 mgr.x (mgr.14152) 292 : cluster [DBG] pgmap v230: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:25.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:24 vm02 bash[17657]: cluster 2026-03-21T06:54:23.588680+0000 mgr.x (mgr.14152) 292 : cluster [DBG] pgmap v230: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:25.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:24 vm02 bash[17657]: cluster 2026-03-21T06:54:23.588680+0000 mgr.x (mgr.14152) 292 : cluster [DBG] pgmap v230: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:25.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:24 vm07 bash[19945]: cluster 2026-03-21T06:54:23.588680+0000 mgr.x (mgr.14152) 292 : cluster [DBG] pgmap v230: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:25.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:24 vm07 bash[19945]: cluster 2026-03-21T06:54:23.588680+0000 mgr.x (mgr.14152) 292 : cluster [DBG] pgmap v230: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:27.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:26 vm04 bash[20194]: cluster 2026-03-21T06:54:25.588965+0000 mgr.x (mgr.14152) 293 : cluster [DBG] pgmap v231: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:27.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:26 vm04 bash[20194]: cluster 2026-03-21T06:54:25.588965+0000 mgr.x (mgr.14152) 293 : cluster [DBG] pgmap v231: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:27.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:26 vm02 bash[17657]: cluster 2026-03-21T06:54:25.588965+0000 mgr.x (mgr.14152) 293 : cluster [DBG] pgmap v231: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:27.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:26 vm02 bash[17657]: cluster 2026-03-21T06:54:25.588965+0000 mgr.x (mgr.14152) 293 : cluster [DBG] pgmap v231: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:26 vm07 bash[19945]: cluster 2026-03-21T06:54:25.588965+0000 mgr.x (mgr.14152) 293 : cluster [DBG] pgmap v231: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:26 vm07 bash[19945]: cluster 2026-03-21T06:54:25.588965+0000 mgr.x (mgr.14152) 293 : cluster [DBG] pgmap v231: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:28 vm04 bash[20194]: cluster 2026-03-21T06:54:27.589267+0000 mgr.x (mgr.14152) 294 : cluster [DBG] pgmap v232: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:28 vm04 bash[20194]: cluster 2026-03-21T06:54:27.589267+0000 mgr.x (mgr.14152) 294 : cluster [DBG] pgmap v232: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:29.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:28 vm02 bash[17657]: cluster 2026-03-21T06:54:27.589267+0000 mgr.x (mgr.14152) 294 : cluster [DBG] pgmap v232: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:29.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:28 vm02 bash[17657]: cluster 2026-03-21T06:54:27.589267+0000 mgr.x (mgr.14152) 294 : cluster [DBG] pgmap v232: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:28 vm07 bash[19945]: cluster 2026-03-21T06:54:27.589267+0000 mgr.x (mgr.14152) 294 : cluster [DBG] pgmap v232: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:28 vm07 bash[19945]: cluster 2026-03-21T06:54:27.589267+0000 mgr.x (mgr.14152) 294 : cluster [DBG] pgmap v232: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:31.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:30 vm04 bash[20194]: cluster 2026-03-21T06:54:29.589649+0000 mgr.x (mgr.14152) 295 : cluster [DBG] pgmap v233: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:31.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:30 vm04 bash[20194]: cluster 2026-03-21T06:54:29.589649+0000 mgr.x (mgr.14152) 295 : cluster [DBG] pgmap v233: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:31.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:30 vm02 bash[17657]: cluster 2026-03-21T06:54:29.589649+0000 mgr.x (mgr.14152) 295 : cluster [DBG] pgmap v233: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:31.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:30 vm02 bash[17657]: cluster 2026-03-21T06:54:29.589649+0000 mgr.x (mgr.14152) 295 : cluster [DBG] pgmap v233: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:30 vm07 bash[19945]: cluster 2026-03-21T06:54:29.589649+0000 mgr.x (mgr.14152) 295 : cluster [DBG] pgmap v233: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:30 vm07 bash[19945]: cluster 2026-03-21T06:54:29.589649+0000 mgr.x (mgr.14152) 295 : cluster [DBG] pgmap v233: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:33.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:32 vm04 bash[20194]: cluster 2026-03-21T06:54:31.590008+0000 mgr.x (mgr.14152) 296 : cluster [DBG] pgmap v234: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:33.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:32 vm04 bash[20194]: cluster 2026-03-21T06:54:31.590008+0000 mgr.x (mgr.14152) 296 : cluster [DBG] pgmap v234: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:33.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:32 vm02 bash[17657]: cluster 2026-03-21T06:54:31.590008+0000 mgr.x (mgr.14152) 296 : cluster [DBG] pgmap v234: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:33.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:32 vm02 bash[17657]: cluster 2026-03-21T06:54:31.590008+0000 mgr.x (mgr.14152) 296 : cluster [DBG] pgmap v234: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:32 vm07 bash[19945]: cluster 2026-03-21T06:54:31.590008+0000 mgr.x (mgr.14152) 296 : cluster [DBG] pgmap v234: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:32 vm07 bash[19945]: cluster 2026-03-21T06:54:31.590008+0000 mgr.x (mgr.14152) 296 : cluster [DBG] pgmap v234: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:35.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:34 vm04 bash[20194]: cluster 2026-03-21T06:54:33.590309+0000 mgr.x (mgr.14152) 297 : cluster [DBG] pgmap v235: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:35.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:34 vm04 bash[20194]: cluster 2026-03-21T06:54:33.590309+0000 mgr.x (mgr.14152) 297 : cluster [DBG] pgmap v235: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:35.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:34 vm02 bash[17657]: cluster 2026-03-21T06:54:33.590309+0000 mgr.x (mgr.14152) 297 : cluster [DBG] pgmap v235: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:35.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:34 vm02 bash[17657]: cluster 2026-03-21T06:54:33.590309+0000 mgr.x (mgr.14152) 297 : cluster [DBG] pgmap v235: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:35.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:34 vm07 bash[19945]: cluster 2026-03-21T06:54:33.590309+0000 mgr.x (mgr.14152) 297 : cluster [DBG] pgmap v235: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:35.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:34 vm07 bash[19945]: cluster 2026-03-21T06:54:33.590309+0000 mgr.x (mgr.14152) 297 : cluster [DBG] pgmap v235: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:36.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:35 vm04 bash[20194]: cluster 2026-03-21T06:54:35.590613+0000 mgr.x (mgr.14152) 298 : cluster [DBG] pgmap v236: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:36.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:35 vm04 bash[20194]: cluster 2026-03-21T06:54:35.590613+0000 mgr.x (mgr.14152) 298 : cluster [DBG] pgmap v236: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:36.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:35 vm02 bash[17657]: cluster 2026-03-21T06:54:35.590613+0000 mgr.x (mgr.14152) 298 : cluster [DBG] pgmap v236: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:36.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:35 vm02 bash[17657]: cluster 2026-03-21T06:54:35.590613+0000 mgr.x (mgr.14152) 298 : cluster [DBG] pgmap v236: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:36.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:35 vm07 bash[19945]: cluster 2026-03-21T06:54:35.590613+0000 mgr.x (mgr.14152) 298 : cluster [DBG] pgmap v236: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:36.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:35 vm07 bash[19945]: cluster 2026-03-21T06:54:35.590613+0000 mgr.x (mgr.14152) 298 : cluster [DBG] pgmap v236: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:38 vm04 bash[20194]: cluster 2026-03-21T06:54:37.590982+0000 mgr.x (mgr.14152) 299 : cluster [DBG] pgmap v237: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:39.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:38 vm04 bash[20194]: cluster 2026-03-21T06:54:37.590982+0000 mgr.x (mgr.14152) 299 : cluster [DBG] pgmap v237: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:39.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:38 vm02 bash[17657]: cluster 2026-03-21T06:54:37.590982+0000 mgr.x (mgr.14152) 299 : cluster [DBG] pgmap v237: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:39.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:38 vm02 bash[17657]: cluster 2026-03-21T06:54:37.590982+0000 mgr.x (mgr.14152) 299 : cluster [DBG] pgmap v237: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:38 vm07 bash[19945]: cluster 2026-03-21T06:54:37.590982+0000 mgr.x (mgr.14152) 299 : cluster [DBG] pgmap v237: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:38 vm07 bash[19945]: cluster 2026-03-21T06:54:37.590982+0000 mgr.x (mgr.14152) 299 : cluster [DBG] pgmap v237: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:41.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:40 vm04 bash[20194]: cluster 2026-03-21T06:54:39.591353+0000 mgr.x (mgr.14152) 300 : cluster [DBG] pgmap v238: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:41.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:40 vm04 bash[20194]: cluster 2026-03-21T06:54:39.591353+0000 mgr.x (mgr.14152) 300 : cluster [DBG] pgmap v238: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:41.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:40 vm02 bash[17657]: cluster 2026-03-21T06:54:39.591353+0000 mgr.x (mgr.14152) 300 : cluster [DBG] pgmap v238: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:41.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:40 vm02 bash[17657]: cluster 2026-03-21T06:54:39.591353+0000 mgr.x (mgr.14152) 300 : cluster [DBG] pgmap v238: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:40 vm07 bash[19945]: cluster 2026-03-21T06:54:39.591353+0000 mgr.x (mgr.14152) 300 : cluster [DBG] pgmap v238: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:40 vm07 bash[19945]: cluster 2026-03-21T06:54:39.591353+0000 mgr.x (mgr.14152) 300 : cluster [DBG] pgmap v238: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:43.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:42 vm04 bash[20194]: cluster 2026-03-21T06:54:41.591725+0000 mgr.x (mgr.14152) 301 : cluster [DBG] pgmap v239: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:43.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:42 vm04 bash[20194]: cluster 2026-03-21T06:54:41.591725+0000 mgr.x (mgr.14152) 301 : cluster [DBG] pgmap v239: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:43.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:42 vm02 bash[17657]: cluster 2026-03-21T06:54:41.591725+0000 mgr.x (mgr.14152) 301 : cluster [DBG] pgmap v239: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:43.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:42 vm02 bash[17657]: cluster 2026-03-21T06:54:41.591725+0000 mgr.x (mgr.14152) 301 : cluster [DBG] pgmap v239: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:42 vm07 bash[19945]: cluster 2026-03-21T06:54:41.591725+0000 mgr.x (mgr.14152) 301 : cluster [DBG] pgmap v239: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:42 vm07 bash[19945]: cluster 2026-03-21T06:54:41.591725+0000 mgr.x (mgr.14152) 301 : cluster [DBG] pgmap v239: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:45.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:44 vm04 bash[20194]: cluster 2026-03-21T06:54:43.592085+0000 mgr.x (mgr.14152) 302 : cluster [DBG] pgmap v240: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:45.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:44 vm04 bash[20194]: cluster 2026-03-21T06:54:43.592085+0000 mgr.x (mgr.14152) 302 : cluster [DBG] pgmap v240: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:45.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:44 vm02 bash[17657]: cluster 2026-03-21T06:54:43.592085+0000 mgr.x (mgr.14152) 302 : cluster [DBG] pgmap v240: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:45.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:44 vm02 bash[17657]: cluster 2026-03-21T06:54:43.592085+0000 mgr.x (mgr.14152) 302 : cluster [DBG] pgmap v240: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:45.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:44 vm07 bash[19945]: cluster 2026-03-21T06:54:43.592085+0000 mgr.x (mgr.14152) 302 : cluster [DBG] pgmap v240: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:45.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:44 vm07 bash[19945]: cluster 2026-03-21T06:54:43.592085+0000 mgr.x (mgr.14152) 302 : cluster [DBG] pgmap v240: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:46 vm04 bash[20194]: cluster 2026-03-21T06:54:45.592456+0000 mgr.x (mgr.14152) 303 : cluster [DBG] pgmap v241: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:46 vm04 bash[20194]: cluster 2026-03-21T06:54:45.592456+0000 mgr.x (mgr.14152) 303 : cluster [DBG] pgmap v241: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:47.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:46 vm02 bash[17657]: cluster 2026-03-21T06:54:45.592456+0000 mgr.x (mgr.14152) 303 : cluster [DBG] pgmap v241: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:47.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:46 vm02 bash[17657]: cluster 2026-03-21T06:54:45.592456+0000 mgr.x (mgr.14152) 303 : cluster [DBG] pgmap v241: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:47.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:46 vm07 bash[19945]: cluster 2026-03-21T06:54:45.592456+0000 mgr.x (mgr.14152) 303 : cluster [DBG] pgmap v241: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:47.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:46 vm07 bash[19945]: cluster 2026-03-21T06:54:45.592456+0000 mgr.x (mgr.14152) 303 : cluster [DBG] pgmap v241: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:49.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:48 vm04 bash[20194]: cluster 2026-03-21T06:54:47.592760+0000 mgr.x (mgr.14152) 304 : cluster [DBG] pgmap v242: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:49.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:48 vm04 bash[20194]: cluster 2026-03-21T06:54:47.592760+0000 mgr.x (mgr.14152) 304 : cluster [DBG] pgmap v242: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:49.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:48 vm02 bash[17657]: cluster 2026-03-21T06:54:47.592760+0000 mgr.x (mgr.14152) 304 : cluster [DBG] pgmap v242: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:49.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:48 vm02 bash[17657]: cluster 2026-03-21T06:54:47.592760+0000 mgr.x (mgr.14152) 304 : cluster [DBG] pgmap v242: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:49.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:48 vm07 bash[19945]: cluster 2026-03-21T06:54:47.592760+0000 mgr.x (mgr.14152) 304 : cluster [DBG] pgmap v242: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:49.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:48 vm07 bash[19945]: cluster 2026-03-21T06:54:47.592760+0000 mgr.x (mgr.14152) 304 : cluster [DBG] pgmap v242: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:51.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:50 vm04 bash[20194]: cluster 2026-03-21T06:54:49.593132+0000 mgr.x (mgr.14152) 305 : cluster [DBG] pgmap v243: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:51.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:50 vm04 bash[20194]: cluster 2026-03-21T06:54:49.593132+0000 mgr.x (mgr.14152) 305 : cluster [DBG] pgmap v243: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:51.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:50 vm02 bash[17657]: cluster 2026-03-21T06:54:49.593132+0000 mgr.x (mgr.14152) 305 : cluster [DBG] pgmap v243: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:51.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:50 vm02 bash[17657]: cluster 2026-03-21T06:54:49.593132+0000 mgr.x (mgr.14152) 305 : cluster [DBG] pgmap v243: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:50 vm07 bash[19945]: cluster 2026-03-21T06:54:49.593132+0000 mgr.x (mgr.14152) 305 : cluster [DBG] pgmap v243: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:50 vm07 bash[19945]: cluster 2026-03-21T06:54:49.593132+0000 mgr.x (mgr.14152) 305 : cluster [DBG] pgmap v243: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:53.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:52 vm04 bash[20194]: cluster 2026-03-21T06:54:51.593455+0000 mgr.x (mgr.14152) 306 : cluster [DBG] pgmap v244: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:53.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:52 vm04 bash[20194]: cluster 2026-03-21T06:54:51.593455+0000 mgr.x (mgr.14152) 306 : cluster [DBG] pgmap v244: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:53.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:52 vm02 bash[17657]: cluster 2026-03-21T06:54:51.593455+0000 mgr.x (mgr.14152) 306 : cluster [DBG] pgmap v244: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:53.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:52 vm02 bash[17657]: cluster 2026-03-21T06:54:51.593455+0000 mgr.x (mgr.14152) 306 : cluster [DBG] pgmap v244: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:52 vm07 bash[19945]: cluster 2026-03-21T06:54:51.593455+0000 mgr.x (mgr.14152) 306 : cluster [DBG] pgmap v244: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:52 vm07 bash[19945]: cluster 2026-03-21T06:54:51.593455+0000 mgr.x (mgr.14152) 306 : cluster [DBG] pgmap v244: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:55.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:54 vm04 bash[20194]: cluster 2026-03-21T06:54:53.593751+0000 mgr.x (mgr.14152) 307 : cluster [DBG] pgmap v245: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:55.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:54 vm04 bash[20194]: cluster 2026-03-21T06:54:53.593751+0000 mgr.x (mgr.14152) 307 : cluster [DBG] pgmap v245: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:54 vm02 bash[17657]: cluster 2026-03-21T06:54:53.593751+0000 mgr.x (mgr.14152) 307 : cluster [DBG] pgmap v245: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:54 vm02 bash[17657]: cluster 2026-03-21T06:54:53.593751+0000 mgr.x (mgr.14152) 307 : cluster [DBG] pgmap v245: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:55.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:54 vm07 bash[19945]: cluster 2026-03-21T06:54:53.593751+0000 mgr.x (mgr.14152) 307 : cluster [DBG] pgmap v245: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:55.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:54 vm07 bash[19945]: cluster 2026-03-21T06:54:53.593751+0000 mgr.x (mgr.14152) 307 : cluster [DBG] pgmap v245: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:56.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:55 vm04 bash[20194]: cluster 2026-03-21T06:54:55.594065+0000 mgr.x (mgr.14152) 308 : cluster [DBG] pgmap v246: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:56.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:55 vm04 bash[20194]: cluster 2026-03-21T06:54:55.594065+0000 mgr.x (mgr.14152) 308 : cluster [DBG] pgmap v246: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:56.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:55 vm02 bash[17657]: cluster 2026-03-21T06:54:55.594065+0000 mgr.x (mgr.14152) 308 : cluster [DBG] pgmap v246: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:56.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:55 vm02 bash[17657]: cluster 2026-03-21T06:54:55.594065+0000 mgr.x (mgr.14152) 308 : cluster [DBG] pgmap v246: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:56.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:55 vm07 bash[19945]: cluster 2026-03-21T06:54:55.594065+0000 mgr.x (mgr.14152) 308 : cluster [DBG] pgmap v246: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:56.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:55 vm07 bash[19945]: cluster 2026-03-21T06:54:55.594065+0000 mgr.x (mgr.14152) 308 : cluster [DBG] pgmap v246: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:58 vm04 bash[20194]: cluster 2026-03-21T06:54:57.594499+0000 mgr.x (mgr.14152) 309 : cluster [DBG] pgmap v247: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:54:58 vm04 bash[20194]: cluster 2026-03-21T06:54:57.594499+0000 mgr.x (mgr.14152) 309 : cluster [DBG] pgmap v247: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:59.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:58 vm02 bash[17657]: cluster 2026-03-21T06:54:57.594499+0000 mgr.x (mgr.14152) 309 : cluster [DBG] pgmap v247: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:59.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:54:58 vm02 bash[17657]: cluster 2026-03-21T06:54:57.594499+0000 mgr.x (mgr.14152) 309 : cluster [DBG] pgmap v247: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:58 vm07 bash[19945]: cluster 2026-03-21T06:54:57.594499+0000 mgr.x (mgr.14152) 309 : cluster [DBG] pgmap v247: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:54:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:54:58 vm07 bash[19945]: cluster 2026-03-21T06:54:57.594499+0000 mgr.x (mgr.14152) 309 : cluster [DBG] pgmap v247: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:01.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:00 vm04 bash[20194]: cluster 2026-03-21T06:54:59.594796+0000 mgr.x (mgr.14152) 310 : cluster [DBG] pgmap v248: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:01.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:00 vm04 bash[20194]: cluster 2026-03-21T06:54:59.594796+0000 mgr.x (mgr.14152) 310 : cluster [DBG] pgmap v248: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:01.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:00 vm02 bash[17657]: cluster 2026-03-21T06:54:59.594796+0000 mgr.x (mgr.14152) 310 : cluster [DBG] pgmap v248: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:01.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:00 vm02 bash[17657]: cluster 2026-03-21T06:54:59.594796+0000 mgr.x (mgr.14152) 310 : cluster [DBG] pgmap v248: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:01.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:00 vm07 bash[19945]: cluster 2026-03-21T06:54:59.594796+0000 mgr.x (mgr.14152) 310 : cluster [DBG] pgmap v248: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:01.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:00 vm07 bash[19945]: cluster 2026-03-21T06:54:59.594796+0000 mgr.x (mgr.14152) 310 : cluster [DBG] pgmap v248: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:02 vm04 bash[20194]: cluster 2026-03-21T06:55:01.595125+0000 mgr.x (mgr.14152) 311 : cluster [DBG] pgmap v249: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:02 vm04 bash[20194]: cluster 2026-03-21T06:55:01.595125+0000 mgr.x (mgr.14152) 311 : cluster [DBG] pgmap v249: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:03.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:02 vm02 bash[17657]: cluster 2026-03-21T06:55:01.595125+0000 mgr.x (mgr.14152) 311 : cluster [DBG] pgmap v249: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:03.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:02 vm02 bash[17657]: cluster 2026-03-21T06:55:01.595125+0000 mgr.x (mgr.14152) 311 : cluster [DBG] pgmap v249: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:03.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:02 vm07 bash[19945]: cluster 2026-03-21T06:55:01.595125+0000 mgr.x (mgr.14152) 311 : cluster [DBG] pgmap v249: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:03.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:02 vm07 bash[19945]: cluster 2026-03-21T06:55:01.595125+0000 mgr.x (mgr.14152) 311 : cluster [DBG] pgmap v249: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:04 vm04 bash[20194]: cluster 2026-03-21T06:55:03.595547+0000 mgr.x (mgr.14152) 312 : cluster [DBG] pgmap v250: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:04 vm04 bash[20194]: cluster 2026-03-21T06:55:03.595547+0000 mgr.x (mgr.14152) 312 : cluster [DBG] pgmap v250: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:05.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:04 vm02 bash[17657]: cluster 2026-03-21T06:55:03.595547+0000 mgr.x (mgr.14152) 312 : cluster [DBG] pgmap v250: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:05.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:04 vm02 bash[17657]: cluster 2026-03-21T06:55:03.595547+0000 mgr.x (mgr.14152) 312 : cluster [DBG] pgmap v250: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:05.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:04 vm07 bash[19945]: cluster 2026-03-21T06:55:03.595547+0000 mgr.x (mgr.14152) 312 : cluster [DBG] pgmap v250: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:05.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:04 vm07 bash[19945]: cluster 2026-03-21T06:55:03.595547+0000 mgr.x (mgr.14152) 312 : cluster [DBG] pgmap v250: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:07.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:06 vm04 bash[20194]: cluster 2026-03-21T06:55:05.595981+0000 mgr.x (mgr.14152) 313 : cluster [DBG] pgmap v251: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:07.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:06 vm04 bash[20194]: cluster 2026-03-21T06:55:05.595981+0000 mgr.x (mgr.14152) 313 : cluster [DBG] pgmap v251: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:07.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:06 vm02 bash[17657]: cluster 2026-03-21T06:55:05.595981+0000 mgr.x (mgr.14152) 313 : cluster [DBG] pgmap v251: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:07.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:06 vm02 bash[17657]: cluster 2026-03-21T06:55:05.595981+0000 mgr.x (mgr.14152) 313 : cluster [DBG] pgmap v251: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:06 vm07 bash[19945]: cluster 2026-03-21T06:55:05.595981+0000 mgr.x (mgr.14152) 313 : cluster [DBG] pgmap v251: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:06 vm07 bash[19945]: cluster 2026-03-21T06:55:05.595981+0000 mgr.x (mgr.14152) 313 : cluster [DBG] pgmap v251: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:08 vm04 bash[20194]: cluster 2026-03-21T06:55:07.596437+0000 mgr.x (mgr.14152) 314 : cluster [DBG] pgmap v252: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:08 vm04 bash[20194]: cluster 2026-03-21T06:55:07.596437+0000 mgr.x (mgr.14152) 314 : cluster [DBG] pgmap v252: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:09.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:08 vm02 bash[17657]: cluster 2026-03-21T06:55:07.596437+0000 mgr.x (mgr.14152) 314 : cluster [DBG] pgmap v252: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:09.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:08 vm02 bash[17657]: cluster 2026-03-21T06:55:07.596437+0000 mgr.x (mgr.14152) 314 : cluster [DBG] pgmap v252: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:08 vm07 bash[19945]: cluster 2026-03-21T06:55:07.596437+0000 mgr.x (mgr.14152) 314 : cluster [DBG] pgmap v252: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:09.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:08 vm07 bash[19945]: cluster 2026-03-21T06:55:07.596437+0000 mgr.x (mgr.14152) 314 : cluster [DBG] pgmap v252: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:11.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:10 vm04 bash[20194]: cluster 2026-03-21T06:55:09.596887+0000 mgr.x (mgr.14152) 315 : cluster [DBG] pgmap v253: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:11.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:10 vm04 bash[20194]: cluster 2026-03-21T06:55:09.596887+0000 mgr.x (mgr.14152) 315 : cluster [DBG] pgmap v253: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:11.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:10 vm02 bash[17657]: cluster 2026-03-21T06:55:09.596887+0000 mgr.x (mgr.14152) 315 : cluster [DBG] pgmap v253: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:11.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:10 vm02 bash[17657]: cluster 2026-03-21T06:55:09.596887+0000 mgr.x (mgr.14152) 315 : cluster [DBG] pgmap v253: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:10 vm07 bash[19945]: cluster 2026-03-21T06:55:09.596887+0000 mgr.x (mgr.14152) 315 : cluster [DBG] pgmap v253: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:10 vm07 bash[19945]: cluster 2026-03-21T06:55:09.596887+0000 mgr.x (mgr.14152) 315 : cluster [DBG] pgmap v253: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:13.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:12 vm04 bash[20194]: cluster 2026-03-21T06:55:11.597261+0000 mgr.x (mgr.14152) 316 : cluster [DBG] pgmap v254: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:13.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:12 vm04 bash[20194]: cluster 2026-03-21T06:55:11.597261+0000 mgr.x (mgr.14152) 316 : cluster [DBG] pgmap v254: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:13.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:12 vm02 bash[17657]: cluster 2026-03-21T06:55:11.597261+0000 mgr.x (mgr.14152) 316 : cluster [DBG] pgmap v254: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:13.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:12 vm02 bash[17657]: cluster 2026-03-21T06:55:11.597261+0000 mgr.x (mgr.14152) 316 : cluster [DBG] pgmap v254: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:12 vm07 bash[19945]: cluster 2026-03-21T06:55:11.597261+0000 mgr.x (mgr.14152) 316 : cluster [DBG] pgmap v254: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:12 vm07 bash[19945]: cluster 2026-03-21T06:55:11.597261+0000 mgr.x (mgr.14152) 316 : cluster [DBG] pgmap v254: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:15.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:14 vm04 bash[20194]: cluster 2026-03-21T06:55:13.597533+0000 mgr.x (mgr.14152) 317 : cluster [DBG] pgmap v255: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:15.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:14 vm04 bash[20194]: cluster 2026-03-21T06:55:13.597533+0000 mgr.x (mgr.14152) 317 : cluster [DBG] pgmap v255: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:15.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:14 vm02 bash[17657]: cluster 2026-03-21T06:55:13.597533+0000 mgr.x (mgr.14152) 317 : cluster [DBG] pgmap v255: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:15.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:14 vm02 bash[17657]: cluster 2026-03-21T06:55:13.597533+0000 mgr.x (mgr.14152) 317 : cluster [DBG] pgmap v255: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:14 vm07 bash[19945]: cluster 2026-03-21T06:55:13.597533+0000 mgr.x (mgr.14152) 317 : cluster [DBG] pgmap v255: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:14 vm07 bash[19945]: cluster 2026-03-21T06:55:13.597533+0000 mgr.x (mgr.14152) 317 : cluster [DBG] pgmap v255: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:16.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:15 vm02 bash[17657]: cluster 2026-03-21T06:55:15.597783+0000 mgr.x (mgr.14152) 318 : cluster [DBG] pgmap v256: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:16.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:15 vm02 bash[17657]: cluster 2026-03-21T06:55:15.597783+0000 mgr.x (mgr.14152) 318 : cluster [DBG] pgmap v256: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:16.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:15 vm07 bash[19945]: cluster 2026-03-21T06:55:15.597783+0000 mgr.x (mgr.14152) 318 : cluster [DBG] pgmap v256: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:16.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:15 vm07 bash[19945]: cluster 2026-03-21T06:55:15.597783+0000 mgr.x (mgr.14152) 318 : cluster [DBG] pgmap v256: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:16.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:15 vm04 bash[20194]: cluster 2026-03-21T06:55:15.597783+0000 mgr.x (mgr.14152) 318 : cluster [DBG] pgmap v256: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:16.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:15 vm04 bash[20194]: cluster 2026-03-21T06:55:15.597783+0000 mgr.x (mgr.14152) 318 : cluster [DBG] pgmap v256: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:18 vm04 bash[20194]: cluster 2026-03-21T06:55:17.598061+0000 mgr.x (mgr.14152) 319 : cluster [DBG] pgmap v257: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:18 vm04 bash[20194]: cluster 2026-03-21T06:55:17.598061+0000 mgr.x (mgr.14152) 319 : cluster [DBG] pgmap v257: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:19.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:18 vm02 bash[17657]: cluster 2026-03-21T06:55:17.598061+0000 mgr.x (mgr.14152) 319 : cluster [DBG] pgmap v257: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:19.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:18 vm02 bash[17657]: cluster 2026-03-21T06:55:17.598061+0000 mgr.x (mgr.14152) 319 : cluster [DBG] pgmap v257: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:18 vm07 bash[19945]: cluster 2026-03-21T06:55:17.598061+0000 mgr.x (mgr.14152) 319 : cluster [DBG] pgmap v257: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:18 vm07 bash[19945]: cluster 2026-03-21T06:55:17.598061+0000 mgr.x (mgr.14152) 319 : cluster [DBG] pgmap v257: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:20 vm04 bash[20194]: cluster 2026-03-21T06:55:19.598360+0000 mgr.x (mgr.14152) 320 : cluster [DBG] pgmap v258: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:20 vm04 bash[20194]: cluster 2026-03-21T06:55:19.598360+0000 mgr.x (mgr.14152) 320 : cluster [DBG] pgmap v258: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:21.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:20 vm02 bash[17657]: cluster 2026-03-21T06:55:19.598360+0000 mgr.x (mgr.14152) 320 : cluster [DBG] pgmap v258: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:21.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:20 vm02 bash[17657]: cluster 2026-03-21T06:55:19.598360+0000 mgr.x (mgr.14152) 320 : cluster [DBG] pgmap v258: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:20 vm07 bash[19945]: cluster 2026-03-21T06:55:19.598360+0000 mgr.x (mgr.14152) 320 : cluster [DBG] pgmap v258: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:20 vm07 bash[19945]: cluster 2026-03-21T06:55:19.598360+0000 mgr.x (mgr.14152) 320 : cluster [DBG] pgmap v258: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:23.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:22 vm04 bash[20194]: cluster 2026-03-21T06:55:21.598744+0000 mgr.x (mgr.14152) 321 : cluster [DBG] pgmap v259: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:23.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:22 vm04 bash[20194]: cluster 2026-03-21T06:55:21.598744+0000 mgr.x (mgr.14152) 321 : cluster [DBG] pgmap v259: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:23.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:22 vm02 bash[17657]: cluster 2026-03-21T06:55:21.598744+0000 mgr.x (mgr.14152) 321 : cluster [DBG] pgmap v259: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:23.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:22 vm02 bash[17657]: cluster 2026-03-21T06:55:21.598744+0000 mgr.x (mgr.14152) 321 : cluster [DBG] pgmap v259: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:22 vm07 bash[19945]: cluster 2026-03-21T06:55:21.598744+0000 mgr.x (mgr.14152) 321 : cluster [DBG] pgmap v259: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:22 vm07 bash[19945]: cluster 2026-03-21T06:55:21.598744+0000 mgr.x (mgr.14152) 321 : cluster [DBG] pgmap v259: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:25.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:24 vm04 bash[20194]: cluster 2026-03-21T06:55:23.599122+0000 mgr.x (mgr.14152) 322 : cluster [DBG] pgmap v260: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:25.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:24 vm04 bash[20194]: cluster 2026-03-21T06:55:23.599122+0000 mgr.x (mgr.14152) 322 : cluster [DBG] pgmap v260: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:25.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:24 vm02 bash[17657]: cluster 2026-03-21T06:55:23.599122+0000 mgr.x (mgr.14152) 322 : cluster [DBG] pgmap v260: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:25.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:24 vm02 bash[17657]: cluster 2026-03-21T06:55:23.599122+0000 mgr.x (mgr.14152) 322 : cluster [DBG] pgmap v260: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:25.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:24 vm07 bash[19945]: cluster 2026-03-21T06:55:23.599122+0000 mgr.x (mgr.14152) 322 : cluster [DBG] pgmap v260: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:25.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:24 vm07 bash[19945]: cluster 2026-03-21T06:55:23.599122+0000 mgr.x (mgr.14152) 322 : cluster [DBG] pgmap v260: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:27.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:26 vm04 bash[20194]: cluster 2026-03-21T06:55:25.599440+0000 mgr.x (mgr.14152) 323 : cluster [DBG] pgmap v261: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:27.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:26 vm04 bash[20194]: cluster 2026-03-21T06:55:25.599440+0000 mgr.x (mgr.14152) 323 : cluster [DBG] pgmap v261: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:27.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:26 vm02 bash[17657]: cluster 2026-03-21T06:55:25.599440+0000 mgr.x (mgr.14152) 323 : cluster [DBG] pgmap v261: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:27.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:26 vm02 bash[17657]: cluster 2026-03-21T06:55:25.599440+0000 mgr.x (mgr.14152) 323 : cluster [DBG] pgmap v261: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:26 vm07 bash[19945]: cluster 2026-03-21T06:55:25.599440+0000 mgr.x (mgr.14152) 323 : cluster [DBG] pgmap v261: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:26 vm07 bash[19945]: cluster 2026-03-21T06:55:25.599440+0000 mgr.x (mgr.14152) 323 : cluster [DBG] pgmap v261: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:28 vm04 bash[20194]: cluster 2026-03-21T06:55:27.599784+0000 mgr.x (mgr.14152) 324 : cluster [DBG] pgmap v262: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:28 vm04 bash[20194]: cluster 2026-03-21T06:55:27.599784+0000 mgr.x (mgr.14152) 324 : cluster [DBG] pgmap v262: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:29.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:28 vm02 bash[17657]: cluster 2026-03-21T06:55:27.599784+0000 mgr.x (mgr.14152) 324 : cluster [DBG] pgmap v262: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:29.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:28 vm02 bash[17657]: cluster 2026-03-21T06:55:27.599784+0000 mgr.x (mgr.14152) 324 : cluster [DBG] pgmap v262: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:28 vm07 bash[19945]: cluster 2026-03-21T06:55:27.599784+0000 mgr.x (mgr.14152) 324 : cluster [DBG] pgmap v262: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:28 vm07 bash[19945]: cluster 2026-03-21T06:55:27.599784+0000 mgr.x (mgr.14152) 324 : cluster [DBG] pgmap v262: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:31.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:30 vm04 bash[20194]: cluster 2026-03-21T06:55:29.600126+0000 mgr.x (mgr.14152) 325 : cluster [DBG] pgmap v263: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:31.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:30 vm04 bash[20194]: cluster 2026-03-21T06:55:29.600126+0000 mgr.x (mgr.14152) 325 : cluster [DBG] pgmap v263: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:31.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:30 vm02 bash[17657]: cluster 2026-03-21T06:55:29.600126+0000 mgr.x (mgr.14152) 325 : cluster [DBG] pgmap v263: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:31.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:30 vm02 bash[17657]: cluster 2026-03-21T06:55:29.600126+0000 mgr.x (mgr.14152) 325 : cluster [DBG] pgmap v263: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:30 vm07 bash[19945]: cluster 2026-03-21T06:55:29.600126+0000 mgr.x (mgr.14152) 325 : cluster [DBG] pgmap v263: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:30 vm07 bash[19945]: cluster 2026-03-21T06:55:29.600126+0000 mgr.x (mgr.14152) 325 : cluster [DBG] pgmap v263: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:33.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:32 vm04 bash[20194]: cluster 2026-03-21T06:55:31.600496+0000 mgr.x (mgr.14152) 326 : cluster [DBG] pgmap v264: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:33.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:32 vm04 bash[20194]: cluster 2026-03-21T06:55:31.600496+0000 mgr.x (mgr.14152) 326 : cluster [DBG] pgmap v264: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:33.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:32 vm02 bash[17657]: cluster 2026-03-21T06:55:31.600496+0000 mgr.x (mgr.14152) 326 : cluster [DBG] pgmap v264: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:33.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:32 vm02 bash[17657]: cluster 2026-03-21T06:55:31.600496+0000 mgr.x (mgr.14152) 326 : cluster [DBG] pgmap v264: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:32 vm07 bash[19945]: cluster 2026-03-21T06:55:31.600496+0000 mgr.x (mgr.14152) 326 : cluster [DBG] pgmap v264: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:32 vm07 bash[19945]: cluster 2026-03-21T06:55:31.600496+0000 mgr.x (mgr.14152) 326 : cluster [DBG] pgmap v264: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:34.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:33 vm02 bash[17657]: cluster 2026-03-21T06:55:33.600778+0000 mgr.x (mgr.14152) 327 : cluster [DBG] pgmap v265: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:34.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:33 vm02 bash[17657]: cluster 2026-03-21T06:55:33.600778+0000 mgr.x (mgr.14152) 327 : cluster [DBG] pgmap v265: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:34.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:33 vm07 bash[19945]: cluster 2026-03-21T06:55:33.600778+0000 mgr.x (mgr.14152) 327 : cluster [DBG] pgmap v265: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:34.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:33 vm07 bash[19945]: cluster 2026-03-21T06:55:33.600778+0000 mgr.x (mgr.14152) 327 : cluster [DBG] pgmap v265: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:33 vm04 bash[20194]: cluster 2026-03-21T06:55:33.600778+0000 mgr.x (mgr.14152) 327 : cluster [DBG] pgmap v265: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:33 vm04 bash[20194]: cluster 2026-03-21T06:55:33.600778+0000 mgr.x (mgr.14152) 327 : cluster [DBG] pgmap v265: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:37.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:36 vm04 bash[20194]: cluster 2026-03-21T06:55:35.601060+0000 mgr.x (mgr.14152) 328 : cluster [DBG] pgmap v266: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:37.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:36 vm04 bash[20194]: cluster 2026-03-21T06:55:35.601060+0000 mgr.x (mgr.14152) 328 : cluster [DBG] pgmap v266: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:37.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:36 vm02 bash[17657]: cluster 2026-03-21T06:55:35.601060+0000 mgr.x (mgr.14152) 328 : cluster [DBG] pgmap v266: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:37.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:36 vm02 bash[17657]: cluster 2026-03-21T06:55:35.601060+0000 mgr.x (mgr.14152) 328 : cluster [DBG] pgmap v266: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:36 vm07 bash[19945]: cluster 2026-03-21T06:55:35.601060+0000 mgr.x (mgr.14152) 328 : cluster [DBG] pgmap v266: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:36 vm07 bash[19945]: cluster 2026-03-21T06:55:35.601060+0000 mgr.x (mgr.14152) 328 : cluster [DBG] pgmap v266: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:38 vm04 bash[20194]: cluster 2026-03-21T06:55:37.601383+0000 mgr.x (mgr.14152) 329 : cluster [DBG] pgmap v267: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:38 vm04 bash[20194]: cluster 2026-03-21T06:55:37.601383+0000 mgr.x (mgr.14152) 329 : cluster [DBG] pgmap v267: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:39.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:38 vm02 bash[17657]: cluster 2026-03-21T06:55:37.601383+0000 mgr.x (mgr.14152) 329 : cluster [DBG] pgmap v267: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:39.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:38 vm02 bash[17657]: cluster 2026-03-21T06:55:37.601383+0000 mgr.x (mgr.14152) 329 : cluster [DBG] pgmap v267: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:38 vm07 bash[19945]: cluster 2026-03-21T06:55:37.601383+0000 mgr.x (mgr.14152) 329 : cluster [DBG] pgmap v267: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:38 vm07 bash[19945]: cluster 2026-03-21T06:55:37.601383+0000 mgr.x (mgr.14152) 329 : cluster [DBG] pgmap v267: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:41.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:40 vm04 bash[20194]: cluster 2026-03-21T06:55:39.601678+0000 mgr.x (mgr.14152) 330 : cluster [DBG] pgmap v268: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:41.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:40 vm04 bash[20194]: cluster 2026-03-21T06:55:39.601678+0000 mgr.x (mgr.14152) 330 : cluster [DBG] pgmap v268: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:41.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:40 vm02 bash[17657]: cluster 2026-03-21T06:55:39.601678+0000 mgr.x (mgr.14152) 330 : cluster [DBG] pgmap v268: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:41.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:40 vm02 bash[17657]: cluster 2026-03-21T06:55:39.601678+0000 mgr.x (mgr.14152) 330 : cluster [DBG] pgmap v268: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:40 vm07 bash[19945]: cluster 2026-03-21T06:55:39.601678+0000 mgr.x (mgr.14152) 330 : cluster [DBG] pgmap v268: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:40 vm07 bash[19945]: cluster 2026-03-21T06:55:39.601678+0000 mgr.x (mgr.14152) 330 : cluster [DBG] pgmap v268: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:43.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:42 vm04 bash[20194]: cluster 2026-03-21T06:55:41.601960+0000 mgr.x (mgr.14152) 331 : cluster [DBG] pgmap v269: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:43.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:42 vm04 bash[20194]: cluster 2026-03-21T06:55:41.601960+0000 mgr.x (mgr.14152) 331 : cluster [DBG] pgmap v269: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:43.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:42 vm02 bash[17657]: cluster 2026-03-21T06:55:41.601960+0000 mgr.x (mgr.14152) 331 : cluster [DBG] pgmap v269: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:43.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:42 vm02 bash[17657]: cluster 2026-03-21T06:55:41.601960+0000 mgr.x (mgr.14152) 331 : cluster [DBG] pgmap v269: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:42 vm07 bash[19945]: cluster 2026-03-21T06:55:41.601960+0000 mgr.x (mgr.14152) 331 : cluster [DBG] pgmap v269: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:42 vm07 bash[19945]: cluster 2026-03-21T06:55:41.601960+0000 mgr.x (mgr.14152) 331 : cluster [DBG] pgmap v269: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:45.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:44 vm04 bash[20194]: cluster 2026-03-21T06:55:43.602257+0000 mgr.x (mgr.14152) 332 : cluster [DBG] pgmap v270: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:45.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:44 vm04 bash[20194]: cluster 2026-03-21T06:55:43.602257+0000 mgr.x (mgr.14152) 332 : cluster [DBG] pgmap v270: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:45.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:44 vm02 bash[17657]: cluster 2026-03-21T06:55:43.602257+0000 mgr.x (mgr.14152) 332 : cluster [DBG] pgmap v270: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:45.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:44 vm02 bash[17657]: cluster 2026-03-21T06:55:43.602257+0000 mgr.x (mgr.14152) 332 : cluster [DBG] pgmap v270: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:45.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:44 vm07 bash[19945]: cluster 2026-03-21T06:55:43.602257+0000 mgr.x (mgr.14152) 332 : cluster [DBG] pgmap v270: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:45.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:44 vm07 bash[19945]: cluster 2026-03-21T06:55:43.602257+0000 mgr.x (mgr.14152) 332 : cluster [DBG] pgmap v270: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:45.684 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:45 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:45.684 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:55:45 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:45.685 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:55:45 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:46.822 INFO:teuthology.orchestra.run.vm07.stdout:Created osd(s) 6,7 on host 'vm07' 2026-03-21T06:55:46.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:46 vm02 bash[17657]: cluster 2026-03-21T06:55:45.602515+0000 mgr.x (mgr.14152) 333 : cluster [DBG] pgmap v271: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:46.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:46 vm02 bash[17657]: cluster 2026-03-21T06:55:45.602515+0000 mgr.x (mgr.14152) 333 : cluster [DBG] pgmap v271: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:46.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:46 vm02 bash[17657]: audit 2026-03-21T06:55:45.619557+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:46.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:46 vm02 bash[17657]: audit 2026-03-21T06:55:45.619557+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:46.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:46 vm02 bash[17657]: audit 2026-03-21T06:55:45.623867+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:46.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:46 vm02 bash[17657]: audit 2026-03-21T06:55:45.623867+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:46.897 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:46 vm07 bash[19945]: cluster 2026-03-21T06:55:45.602515+0000 mgr.x (mgr.14152) 333 : cluster [DBG] pgmap v271: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:46.897 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:46 vm07 bash[19945]: cluster 2026-03-21T06:55:45.602515+0000 mgr.x (mgr.14152) 333 : cluster [DBG] pgmap v271: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:46.897 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:46 vm07 bash[19945]: audit 2026-03-21T06:55:45.619557+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:46.897 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:46 vm07 bash[19945]: audit 2026-03-21T06:55:45.619557+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:46.897 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:46 vm07 bash[19945]: audit 2026-03-21T06:55:45.623867+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:46.897 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:46 vm07 bash[19945]: audit 2026-03-21T06:55:45.623867+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:46.930 DEBUG:teuthology.orchestra.run.vm07:osd.7> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.7.service 2026-03-21T06:55:46.931 INFO:tasks.cephadm:Waiting for 8 OSDs to come up... 2026-03-21T06:55:46.931 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd stat -f json 2026-03-21T06:55:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:46 vm04 bash[20194]: cluster 2026-03-21T06:55:45.602515+0000 mgr.x (mgr.14152) 333 : cluster [DBG] pgmap v271: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:46 vm04 bash[20194]: cluster 2026-03-21T06:55:45.602515+0000 mgr.x (mgr.14152) 333 : cluster [DBG] pgmap v271: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:46 vm04 bash[20194]: audit 2026-03-21T06:55:45.619557+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:46 vm04 bash[20194]: audit 2026-03-21T06:55:45.619557+0000 mon.a (mon.0) 698 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:46 vm04 bash[20194]: audit 2026-03-21T06:55:45.623867+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:46 vm04 bash[20194]: audit 2026-03-21T06:55:45.623867+0000 mon.a (mon.0) 699 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.151 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:46 vm07 bash[37095]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-7 2026-03-21T06:55:47.151 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:46 vm07 bash[37095]: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph prime-osd-dir --dev /dev/vg_nvme/lv_2 --path /var/lib/ceph/osd/ceph-7 --no-mon-config 2026-03-21T06:55:47.151 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:47 vm07 bash[37095]: Running command: /usr/bin/ln -snf /dev/vg_nvme/lv_2 /var/lib/ceph/osd/ceph-7/block 2026-03-21T06:55:47.151 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:47 vm07 bash[37095]: Running command: /usr/bin/chown -h ceph:ceph /var/lib/ceph/osd/ceph-7/block 2026-03-21T06:55:47.151 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:47 vm07 bash[37095]: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-1 2026-03-21T06:55:47.151 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:47 vm07 bash[37095]: Running command: /usr/bin/chown -R ceph:ceph /var/lib/ceph/osd/ceph-7 2026-03-21T06:55:47.151 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:47 vm07 bash[37095]: --> ceph-volume lvm activate successful for osd ID: 7 2026-03-21T06:55:47.188 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:47.519 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:55:47.580 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":53,"num_osds":8,"num_up_osds":8,"osd_up_since":1774076029,"num_in_osds":8,"osd_in_since":1774076018,"num_remapped_pgs":0} 2026-03-21T06:55:47.580 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd dump --format=json 2026-03-21T06:55:47.629 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:47 vm07 bash[37882]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-7 2026-03-21T06:55:47.629 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:47 vm07 bash[37964]: debug 2026-03-21T06:55:47.394+0000 7f4d1d69d640 1 -- 192.168.123.107:0/1433226323 <== mon.1 v2:192.168.123.107:3300/0 4 ==== auth_reply(proto 2 0 (0) Success) ==== 194+0+0 (secure 0 0 0) 0x55912c651860 con 0x55912c75e000 2026-03-21T06:55:47.828 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: audit 2026-03-21T06:55:46.816820+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: audit 2026-03-21T06:55:46.816820+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: audit 2026-03-21T06:55:46.820572+0000 mon.a (mon.0) 701 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: audit 2026-03-21T06:55:46.820572+0000 mon.a (mon.0) 701 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.410973+0000 mon.a (mon.0) 702 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.410973+0000 mon.a (mon.0) 702 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411025+0000 mon.a (mon.0) 703 : cluster [INF] osd.7 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411025+0000 mon.a (mon.0) 703 : cluster [INF] osd.7 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411156+0000 mon.a (mon.0) 704 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411156+0000 mon.a (mon.0) 704 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411279+0000 mon.a (mon.0) 705 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411279+0000 mon.a (mon.0) 705 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411353+0000 mon.a (mon.0) 706 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411353+0000 mon.a (mon.0) 706 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411444+0000 mon.a (mon.0) 707 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411444+0000 mon.a (mon.0) 707 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411499+0000 mon.a (mon.0) 708 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411499+0000 mon.a (mon.0) 708 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411568+0000 mon.a (mon.0) 709 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411568+0000 mon.a (mon.0) 709 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411640+0000 mon.a (mon.0) 710 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:47.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411640+0000 mon.a (mon.0) 710 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411722+0000 mon.a (mon.0) 711 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411722+0000 mon.a (mon.0) 711 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411796+0000 mon.a (mon.0) 712 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411796+0000 mon.a (mon.0) 712 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411874+0000 mon.a (mon.0) 713 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411874+0000 mon.a (mon.0) 713 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411943+0000 mon.a (mon.0) 714 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.411943+0000 mon.a (mon.0) 714 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.412023+0000 mon.a (mon.0) 715 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.412023+0000 mon.a (mon.0) 715 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.412102+0000 mon.a (mon.0) 716 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.412102+0000 mon.a (mon.0) 716 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.414881+0000 mon.a (mon.0) 717 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.414881+0000 mon.a (mon.0) 717 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.414962+0000 mon.a (mon.0) 718 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.414962+0000 mon.a (mon.0) 718 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: audit 2026-03-21T06:55:47.521205+0000 mon.a (mon.0) 719 : audit [DBG] from='client.? 192.168.123.102:0/2210532337' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: audit 2026-03-21T06:55:47.521205+0000 mon.a (mon.0) 719 : audit [DBG] from='client.? 192.168.123.102:0/2210532337' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.602841+0000 mgr.x (mgr.14152) 334 : cluster [DBG] pgmap v272: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:47.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:47 vm02 bash[17657]: cluster 2026-03-21T06:55:47.602841+0000 mgr.x (mgr.14152) 334 : cluster [DBG] pgmap v272: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: audit 2026-03-21T06:55:46.816820+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: audit 2026-03-21T06:55:46.816820+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: audit 2026-03-21T06:55:46.820572+0000 mon.a (mon.0) 701 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: audit 2026-03-21T06:55:46.820572+0000 mon.a (mon.0) 701 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.410973+0000 mon.a (mon.0) 702 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.410973+0000 mon.a (mon.0) 702 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411025+0000 mon.a (mon.0) 703 : cluster [INF] osd.7 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411025+0000 mon.a (mon.0) 703 : cluster [INF] osd.7 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411156+0000 mon.a (mon.0) 704 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411156+0000 mon.a (mon.0) 704 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411279+0000 mon.a (mon.0) 705 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411279+0000 mon.a (mon.0) 705 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411353+0000 mon.a (mon.0) 706 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411353+0000 mon.a (mon.0) 706 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411444+0000 mon.a (mon.0) 707 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411444+0000 mon.a (mon.0) 707 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411499+0000 mon.a (mon.0) 708 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411499+0000 mon.a (mon.0) 708 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411568+0000 mon.a (mon.0) 709 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411568+0000 mon.a (mon.0) 709 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411640+0000 mon.a (mon.0) 710 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411640+0000 mon.a (mon.0) 710 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411722+0000 mon.a (mon.0) 711 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411722+0000 mon.a (mon.0) 711 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411796+0000 mon.a (mon.0) 712 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411796+0000 mon.a (mon.0) 712 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411874+0000 mon.a (mon.0) 713 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411874+0000 mon.a (mon.0) 713 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411943+0000 mon.a (mon.0) 714 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.411943+0000 mon.a (mon.0) 714 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.412023+0000 mon.a (mon.0) 715 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.412023+0000 mon.a (mon.0) 715 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.412102+0000 mon.a (mon.0) 716 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.412102+0000 mon.a (mon.0) 716 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.414881+0000 mon.a (mon.0) 717 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.414881+0000 mon.a (mon.0) 717 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.414962+0000 mon.a (mon.0) 718 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.414962+0000 mon.a (mon.0) 718 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: audit 2026-03-21T06:55:47.521205+0000 mon.a (mon.0) 719 : audit [DBG] from='client.? 192.168.123.102:0/2210532337' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-21T06:55:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: audit 2026-03-21T06:55:47.521205+0000 mon.a (mon.0) 719 : audit [DBG] from='client.? 192.168.123.102:0/2210532337' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-21T06:55:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.602841+0000 mgr.x (mgr.14152) 334 : cluster [DBG] pgmap v272: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:47 vm07 bash[19945]: cluster 2026-03-21T06:55:47.602841+0000 mgr.x (mgr.14152) 334 : cluster [DBG] pgmap v272: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:47.902 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:47 vm07 bash[37964]: debug 2026-03-21T06:55:47.650+0000 7f4d1ff078c0 -1 Falling back to public interface 2026-03-21T06:55:48.170 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:55:48.170 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":54,"fsid":"b16ecafc-24f1-11f1-8ede-8330751617ee","created":"2026-03-21T06:47:47.924264+0000","modified":"2026-03-21T06:55:47.822170+0000","last_up_change":"2026-03-21T06:55:47.822170+0000","last_in_change":"2026-03-21T06:53:38.245564+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":22,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":1,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-21T06:51:07.780816+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"24","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"5a293892-6a20-454b-9116-106cddf14695","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":8,"last_clean_end":9,"up_from":12,"up_thru":27,"down_at":11,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6803","nonce":2802530539}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6805","nonce":2802530539}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6809","nonce":2802530539}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6807","nonce":2802530539}]},"public_addr":"192.168.123.102:6803/2802530539","cluster_addr":"192.168.123.102:6805/2802530539","heartbeat_back_addr":"192.168.123.102:6809/2802530539","heartbeat_front_addr":"192.168.123.102:6807/2802530539","state":["exists","up"]},{"osd":1,"uuid":"38235c66-ddf2-470c-8f4a-dac1efac5f75","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":16,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6811","nonce":2868931959}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6813","nonce":2868931959}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6817","nonce":2868931959}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6815","nonce":2868931959}]},"public_addr":"192.168.123.102:6811/2868931959","cluster_addr":"192.168.123.102:6813/2868931959","heartbeat_back_addr":"192.168.123.102:6817/2868931959","heartbeat_front_addr":"192.168.123.102:6815/2868931959","state":["exists","up"]},{"osd":2,"uuid":"2bbd253a-cfb7-4d35-a80c-5ab1d023b91e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":22,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6801","nonce":3750025046}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6803","nonce":3750025046}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6807","nonce":3750025046}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6805","nonce":3750025046}]},"public_addr":"192.168.123.104:6801/3750025046","cluster_addr":"192.168.123.104:6803/3750025046","heartbeat_back_addr":"192.168.123.104:6807/3750025046","heartbeat_front_addr":"192.168.123.104:6805/3750025046","state":["exists","up"]},{"osd":3,"uuid":"c2851fc5-1154-4c44-91cd-287a260b97a7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":41,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6808","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6809","nonce":3282486205}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6810","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6811","nonce":3282486205}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6814","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6815","nonce":3282486205}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6812","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6813","nonce":3282486205}]},"public_addr":"192.168.123.104:6809/3282486205","cluster_addr":"192.168.123.104:6811/3282486205","heartbeat_back_addr":"192.168.123.104:6815/3282486205","heartbeat_front_addr":"192.168.123.104:6813/3282486205","state":["exists","up"]},{"osd":4,"uuid":"22bec0cb-fc4e-4173-bab0-1ea7c238683c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6816","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6817","nonce":3736423955}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6818","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6819","nonce":3736423955}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6822","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6823","nonce":3736423955}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6820","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6821","nonce":3736423955}]},"public_addr":"192.168.123.104:6817/3736423955","cluster_addr":"192.168.123.104:6819/3736423955","heartbeat_back_addr":"192.168.123.104:6823/3736423955","heartbeat_front_addr":"192.168.123.104:6821/3736423955","state":["exists","up"]},{"osd":5,"uuid":"a697bb58-dae5-4cba-b085-8e5769734d61","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":36,"last_clean_end":38,"up_from":41,"up_thru":0,"down_at":39,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6800","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6801","nonce":102330835}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6802","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6803","nonce":102330835}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6806","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6807","nonce":102330835}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6804","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6805","nonce":102330835}]},"public_addr":"192.168.123.107:6801/102330835","cluster_addr":"192.168.123.107:6803/102330835","heartbeat_back_addr":"192.168.123.107:6807/102330835","heartbeat_front_addr":"192.168.123.107:6805/102330835","state":["exists","up"]},{"osd":6,"uuid":"7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":45,"last_clean_end":46,"up_from":49,"up_thru":0,"down_at":47,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6808","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6809","nonce":2090727328}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6810","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6811","nonce":2090727328}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6814","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6815","nonce":2090727328}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6812","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6813","nonce":2090727328}]},"public_addr":"192.168.123.107:6809/2090727328","cluster_addr":"192.168.123.107:6811/2090727328","heartbeat_back_addr":"192.168.123.107:6815/2090727328","heartbeat_front_addr":"192.168.123.107:6813/2090727328","state":["exists","up"]},{"osd":7,"uuid":"8e806dd2-e80e-40ab-8996-ff252b7491d9","up":0,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":52,"up_thru":0,"down_at":54,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6816","nonce":3920516005},{"type":"v1","addr":"192.168.123.107:6817","nonce":3920516005}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6818","nonce":3920516005},{"type":"v1","addr":"192.168.123.107:6819","nonce":3920516005}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6822","nonce":3920516005},{"type":"v1","addr":"192.168.123.107:6823","nonce":3920516005}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6820","nonce":3920516005},{"type":"v1","addr":"192.168.123.107:6821","nonce":3920516005}]},"public_addr":"192.168.123.107:6817/3920516005","cluster_addr":"192.168.123.107:6819/3920516005","heartbeat_back_addr":"192.168.123.107:6823/3920516005","heartbeat_front_addr":"192.168.123.107:6821/3920516005","state":["exists"]}],"osd_xinfo":[{"osd":0,"down_stamp":"2026-03-21T06:50:52.082988+0000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:48:49.774464+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:50:58.549660+0000","dead_epoch":15},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:05.687654+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:13.735249+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:20.945258+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:28.648539+0000","dead_epoch":39},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:53:37.274136+0000","dead_epoch":46},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:53:47.648093+0000","dead_epoch":54}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:6801/4266123507":"2026-03-22T06:48:09.528930+0000","192.168.123.102:6800/4266123507":"2026-03-22T06:48:09.528930+0000","192.168.123.102:0/1909593542":"2026-03-22T06:48:09.528930+0000","192.168.123.102:0/1252995422":"2026-03-22T06:48:09.528930+0000","192.168.123.102:0/1580070482":"2026-03-22T06:48:09.528930+0000","192.168.123.102:6801/260656842":"2026-03-22T06:47:58.527875+0000","192.168.123.102:6800/260656842":"2026-03-22T06:47:58.527875+0000","192.168.123.102:0/2023406208":"2026-03-22T06:47:58.527875+0000","192.168.123.102:0/1804345957":"2026-03-22T06:47:58.527875+0000","192.168.123.102:0/288488007":"2026-03-22T06:47:58.527875+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-21T06:55:48.230 INFO:tasks.cephadm.ceph_manager.ceph:[{'pool': 1, 'pool_name': '.mgr', 'create_time': '2026-03-21T06:51:07.780816+0000', 'flags': 1, 'flags_names': 'hashpspool', 'type': 1, 'size': 3, 'min_size': 2, 'crush_rule': 0, 'peering_crush_bucket_count': 0, 'peering_crush_bucket_target': 0, 'peering_crush_bucket_barrier': 0, 'peering_crush_bucket_mandatory_member': 2147483647, 'is_stretch_pool': False, 'object_hash': 2, 'pg_autoscale_mode': 'off', 'pg_num': 1, 'pg_placement_num': 1, 'pg_placement_num_target': 1, 'pg_num_target': 1, 'pg_num_pending': 1, 'last_pg_merge_meta': {'source_pgid': '0.0', 'ready_epoch': 0, 'last_epoch_started': 0, 'last_epoch_clean': 0, 'source_version': "0'0", 'target_version': "0'0"}, 'last_change': '24', 'last_force_op_resend': '0', 'last_force_op_resend_prenautilus': '0', 'last_force_op_resend_preluminous': '0', 'auid': 0, 'snap_mode': 'selfmanaged', 'snap_seq': 0, 'snap_epoch': 0, 'pool_snaps': [], 'removed_snaps': '[]', 'quota_max_bytes': 0, 'quota_max_objects': 0, 'tiers': [], 'tier_of': -1, 'read_tier': -1, 'write_tier': -1, 'cache_mode': 'none', 'target_max_bytes': 0, 'target_max_objects': 0, 'cache_target_dirty_ratio_micro': 400000, 'cache_target_dirty_high_ratio_micro': 600000, 'cache_target_full_ratio_micro': 800000, 'cache_min_flush_age': 0, 'cache_min_evict_age': 0, 'erasure_code_profile': '', 'hit_set_params': {'type': 'none'}, 'hit_set_period': 0, 'hit_set_count': 0, 'use_gmt_hitset': True, 'min_read_recency_for_promote': 0, 'min_write_recency_for_promote': 0, 'hit_set_grade_decay_rate': 0, 'hit_set_search_last_n': 0, 'grade_table': [], 'stripe_width': 0, 'expected_num_objects': 0, 'fast_read': False, 'nonprimary_shards': '{}', 'options': {'pg_num_max': 32, 'pg_num_min': 1}, 'application_metadata': {'mgr': {}}, 'read_balance': {'score_type': 'Fair distribution', 'score_acting': 7.889999866485596, 'score_stable': 7.889999866485596, 'optimal_score': 0.3799999952316284, 'raw_score_acting': 3, 'raw_score_stable': 3, 'primary_affinity_weighted': 1, 'average_primary_affinity': 1, 'average_primary_affinity_weighted': 1}}] 2026-03-21T06:55:48.231 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd pool get .mgr pg_num 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: audit 2026-03-21T06:55:46.816820+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: audit 2026-03-21T06:55:46.816820+0000 mon.a (mon.0) 700 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: audit 2026-03-21T06:55:46.820572+0000 mon.a (mon.0) 701 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: audit 2026-03-21T06:55:46.820572+0000 mon.a (mon.0) 701 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.410973+0000 mon.a (mon.0) 702 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.410973+0000 mon.a (mon.0) 702 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411025+0000 mon.a (mon.0) 703 : cluster [INF] osd.7 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411025+0000 mon.a (mon.0) 703 : cluster [INF] osd.7 failed (root=default,host=vm07) (connection refused reported by osd.0) 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411156+0000 mon.a (mon.0) 704 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411156+0000 mon.a (mon.0) 704 : cluster [DBG] osd.7 reported immediately failed by osd.0 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411279+0000 mon.a (mon.0) 705 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411279+0000 mon.a (mon.0) 705 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411353+0000 mon.a (mon.0) 706 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411353+0000 mon.a (mon.0) 706 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411444+0000 mon.a (mon.0) 707 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411444+0000 mon.a (mon.0) 707 : cluster [DBG] osd.7 reported immediately failed by osd.2 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411499+0000 mon.a (mon.0) 708 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411499+0000 mon.a (mon.0) 708 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411568+0000 mon.a (mon.0) 709 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411568+0000 mon.a (mon.0) 709 : cluster [DBG] osd.7 reported immediately failed by osd.5 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411640+0000 mon.a (mon.0) 710 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411640+0000 mon.a (mon.0) 710 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411722+0000 mon.a (mon.0) 711 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411722+0000 mon.a (mon.0) 711 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411796+0000 mon.a (mon.0) 712 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411796+0000 mon.a (mon.0) 712 : cluster [DBG] osd.7 reported immediately failed by osd.3 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411874+0000 mon.a (mon.0) 713 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411874+0000 mon.a (mon.0) 713 : cluster [DBG] osd.7 reported immediately failed by osd.6 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411943+0000 mon.a (mon.0) 714 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.411943+0000 mon.a (mon.0) 714 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.412023+0000 mon.a (mon.0) 715 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.412023+0000 mon.a (mon.0) 715 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.412102+0000 mon.a (mon.0) 716 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.412102+0000 mon.a (mon.0) 716 : cluster [DBG] osd.7 reported immediately failed by osd.4 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.414881+0000 mon.a (mon.0) 717 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.414881+0000 mon.a (mon.0) 717 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.414962+0000 mon.a (mon.0) 718 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.414962+0000 mon.a (mon.0) 718 : cluster [DBG] osd.7 reported immediately failed by osd.1 2026-03-21T06:55:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: audit 2026-03-21T06:55:47.521205+0000 mon.a (mon.0) 719 : audit [DBG] from='client.? 192.168.123.102:0/2210532337' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-21T06:55:48.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: audit 2026-03-21T06:55:47.521205+0000 mon.a (mon.0) 719 : audit [DBG] from='client.? 192.168.123.102:0/2210532337' entity='client.admin' cmd={"prefix": "osd stat", "format": "json"} : dispatch 2026-03-21T06:55:48.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.602841+0000 mgr.x (mgr.14152) 334 : cluster [DBG] pgmap v272: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:48.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:47 vm04 bash[20194]: cluster 2026-03-21T06:55:47.602841+0000 mgr.x (mgr.14152) 334 : cluster [DBG] pgmap v272: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:48.401 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:47 vm07 bash[37964]: debug 2026-03-21T06:55:47.922+0000 7f4d1ff078c0 -1 osd.7 53 log_to_monitors true 2026-03-21T06:55:48.484 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:48.791 INFO:teuthology.orchestra.run.vm02.stdout:pg_num: 1 2026-03-21T06:55:48.845 INFO:tasks.cephadm:Adding ceph.iscsi.iscsi.a on vm02 2026-03-21T06:55:48.845 INFO:tasks.cephadm:Adding ceph.iscsi.iscsi.b on vm07 2026-03-21T06:55:48.846 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd pool create datapool 3 3 replicated 2026-03-21T06:55:49.078 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: cluster 2026-03-21T06:55:47.822304+0000 mon.a (mon.0) 720 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: cluster 2026-03-21T06:55:47.822304+0000 mon.a (mon.0) 720 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: cluster 2026-03-21T06:55:47.830650+0000 mon.a (mon.0) 721 : cluster [DBG] osdmap e54: 8 total, 7 up, 8 in 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: cluster 2026-03-21T06:55:47.830650+0000 mon.a (mon.0) 721 : cluster [DBG] osdmap e54: 8 total, 7 up, 8 in 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:47.962742+0000 mon.c (mon.1) 24 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:47.962742+0000 mon.c (mon.1) 24 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:47.964405+0000 mon.a (mon.0) 722 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:47.964405+0000 mon.a (mon.0) 722 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:47.992375+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:47.992375+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:47.997226+0000 mon.a (mon.0) 724 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:47.997226+0000 mon.a (mon.0) 724 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:48.013764+0000 mon.a (mon.0) 725 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:48.013764+0000 mon.a (mon.0) 725 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:48.171687+0000 mon.a (mon.0) 726 : audit [DBG] from='client.? 192.168.123.102:0/400200998' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:48.171687+0000 mon.a (mon.0) 726 : audit [DBG] from='client.? 192.168.123.102:0/400200998' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:48.792951+0000 mon.a (mon.0) 727 : audit [DBG] from='client.? 192.168.123.102:0/1636128846' entity='client.admin' cmd={"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"} : dispatch 2026-03-21T06:55:49.091 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:48 vm07 bash[19945]: audit 2026-03-21T06:55:48.792951+0000 mon.a (mon.0) 727 : audit [DBG] from='client.? 192.168.123.102:0/1636128846' entity='client.admin' cmd={"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"} : dispatch 2026-03-21T06:55:49.091 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:48 vm07 bash[37964]: debug 2026-03-21T06:55:48.838+0000 7f4d16c6e640 -1 osd.7 53 set_numa_affinity unable to identify public interface '' numa node: (2) No such file or directory 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: cluster 2026-03-21T06:55:47.822304+0000 mon.a (mon.0) 720 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: cluster 2026-03-21T06:55:47.822304+0000 mon.a (mon.0) 720 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: cluster 2026-03-21T06:55:47.830650+0000 mon.a (mon.0) 721 : cluster [DBG] osdmap e54: 8 total, 7 up, 8 in 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: cluster 2026-03-21T06:55:47.830650+0000 mon.a (mon.0) 721 : cluster [DBG] osdmap e54: 8 total, 7 up, 8 in 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:47.962742+0000 mon.c (mon.1) 24 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:47.962742+0000 mon.c (mon.1) 24 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:47.964405+0000 mon.a (mon.0) 722 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:47.964405+0000 mon.a (mon.0) 722 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:47.992375+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:47.992375+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:47.997226+0000 mon.a (mon.0) 724 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:47.997226+0000 mon.a (mon.0) 724 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:48.013764+0000 mon.a (mon.0) 725 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:48.013764+0000 mon.a (mon.0) 725 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:48.171687+0000 mon.a (mon.0) 726 : audit [DBG] from='client.? 192.168.123.102:0/400200998' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:48.171687+0000 mon.a (mon.0) 726 : audit [DBG] from='client.? 192.168.123.102:0/400200998' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:48.792951+0000 mon.a (mon.0) 727 : audit [DBG] from='client.? 192.168.123.102:0/1636128846' entity='client.admin' cmd={"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"} : dispatch 2026-03-21T06:55:49.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:48 vm02 bash[17657]: audit 2026-03-21T06:55:48.792951+0000 mon.a (mon.0) 727 : audit [DBG] from='client.? 192.168.123.102:0/1636128846' entity='client.admin' cmd={"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"} : dispatch 2026-03-21T06:55:49.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: cluster 2026-03-21T06:55:47.822304+0000 mon.a (mon.0) 720 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: cluster 2026-03-21T06:55:47.822304+0000 mon.a (mon.0) 720 : cluster [WRN] Health check failed: 1 osds down (OSD_DOWN) 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: cluster 2026-03-21T06:55:47.830650+0000 mon.a (mon.0) 721 : cluster [DBG] osdmap e54: 8 total, 7 up, 8 in 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: cluster 2026-03-21T06:55:47.830650+0000 mon.a (mon.0) 721 : cluster [DBG] osdmap e54: 8 total, 7 up, 8 in 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:47.962742+0000 mon.c (mon.1) 24 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:47.962742+0000 mon.c (mon.1) 24 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:47.964405+0000 mon.a (mon.0) 722 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:47.964405+0000 mon.a (mon.0) 722 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]} : dispatch 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:47.992375+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:47.992375+0000 mon.a (mon.0) 723 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:47.997226+0000 mon.a (mon.0) 724 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:47.997226+0000 mon.a (mon.0) 724 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:48.013764+0000 mon.a (mon.0) 725 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:48.013764+0000 mon.a (mon.0) 725 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:48.171687+0000 mon.a (mon.0) 726 : audit [DBG] from='client.? 192.168.123.102:0/400200998' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:48.171687+0000 mon.a (mon.0) 726 : audit [DBG] from='client.? 192.168.123.102:0/400200998' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:48.792951+0000 mon.a (mon.0) 727 : audit [DBG] from='client.? 192.168.123.102:0/1636128846' entity='client.admin' cmd={"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"} : dispatch 2026-03-21T06:55:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:48 vm04 bash[20194]: audit 2026-03-21T06:55:48.792951+0000 mon.a (mon.0) 727 : audit [DBG] from='client.? 192.168.123.102:0/1636128846' entity='client.admin' cmd={"prefix": "osd pool get", "pool": ".mgr", "var": "pg_num"} : dispatch 2026-03-21T06:55:49.843 INFO:teuthology.orchestra.run.vm07.stderr:pool 'datapool' created 2026-03-21T06:55:49.931 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- rbd pool init datapool 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: audit 2026-03-21T06:55:48.834062+0000 mon.a (mon.0) 728 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: audit 2026-03-21T06:55:48.834062+0000 mon.a (mon.0) 728 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: cluster 2026-03-21T06:55:48.836738+0000 mon.a (mon.0) 729 : cluster [DBG] osdmap e55: 8 total, 7 up, 8 in 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: cluster 2026-03-21T06:55:48.836738+0000 mon.a (mon.0) 729 : cluster [DBG] osdmap e55: 8 total, 7 up, 8 in 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: audit 2026-03-21T06:55:48.837180+0000 mon.c (mon.1) 25 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: audit 2026-03-21T06:55:48.837180+0000 mon.c (mon.1) 25 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: audit 2026-03-21T06:55:48.838770+0000 mon.a (mon.0) 730 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: audit 2026-03-21T06:55:48.838770+0000 mon.a (mon.0) 730 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: audit 2026-03-21T06:55:49.415312+0000 mon.c (mon.1) 26 : audit [INF] from='client.? 192.168.123.107:0/3612281354' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: audit 2026-03-21T06:55:49.415312+0000 mon.c (mon.1) 26 : audit [INF] from='client.? 192.168.123.107:0/3612281354' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: audit 2026-03-21T06:55:49.417198+0000 mon.a (mon.0) 731 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: audit 2026-03-21T06:55:49.417198+0000 mon.a (mon.0) 731 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: cluster 2026-03-21T06:55:49.603151+0000 mgr.x (mgr.14152) 335 : cluster [DBG] pgmap v275: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:50.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:49 vm02 bash[17657]: cluster 2026-03-21T06:55:49.603151+0000 mgr.x (mgr.14152) 335 : cluster [DBG] pgmap v275: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: audit 2026-03-21T06:55:48.834062+0000 mon.a (mon.0) 728 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: audit 2026-03-21T06:55:48.834062+0000 mon.a (mon.0) 728 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: cluster 2026-03-21T06:55:48.836738+0000 mon.a (mon.0) 729 : cluster [DBG] osdmap e55: 8 total, 7 up, 8 in 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: cluster 2026-03-21T06:55:48.836738+0000 mon.a (mon.0) 729 : cluster [DBG] osdmap e55: 8 total, 7 up, 8 in 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: audit 2026-03-21T06:55:48.837180+0000 mon.c (mon.1) 25 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: audit 2026-03-21T06:55:48.837180+0000 mon.c (mon.1) 25 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: audit 2026-03-21T06:55:48.838770+0000 mon.a (mon.0) 730 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: audit 2026-03-21T06:55:48.838770+0000 mon.a (mon.0) 730 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: audit 2026-03-21T06:55:49.415312+0000 mon.c (mon.1) 26 : audit [INF] from='client.? 192.168.123.107:0/3612281354' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: audit 2026-03-21T06:55:49.415312+0000 mon.c (mon.1) 26 : audit [INF] from='client.? 192.168.123.107:0/3612281354' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: audit 2026-03-21T06:55:49.417198+0000 mon.a (mon.0) 731 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: audit 2026-03-21T06:55:49.417198+0000 mon.a (mon.0) 731 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: cluster 2026-03-21T06:55:49.603151+0000 mgr.x (mgr.14152) 335 : cluster [DBG] pgmap v275: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:50.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:49 vm07 bash[19945]: cluster 2026-03-21T06:55:49.603151+0000 mgr.x (mgr.14152) 335 : cluster [DBG] pgmap v275: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:50.163 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:55:50.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: audit 2026-03-21T06:55:48.834062+0000 mon.a (mon.0) 728 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: audit 2026-03-21T06:55:48.834062+0000 mon.a (mon.0) 728 : audit [INF] from='osd.7 ' entity='osd.7' cmd='[{"prefix": "osd crush set-device-class", "class": "hdd", "ids": ["7"]}]': finished 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: cluster 2026-03-21T06:55:48.836738+0000 mon.a (mon.0) 729 : cluster [DBG] osdmap e55: 8 total, 7 up, 8 in 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: cluster 2026-03-21T06:55:48.836738+0000 mon.a (mon.0) 729 : cluster [DBG] osdmap e55: 8 total, 7 up, 8 in 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: audit 2026-03-21T06:55:48.837180+0000 mon.c (mon.1) 25 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: audit 2026-03-21T06:55:48.837180+0000 mon.c (mon.1) 25 : audit [INF] from='osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136]' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: audit 2026-03-21T06:55:48.838770+0000 mon.a (mon.0) 730 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: audit 2026-03-21T06:55:48.838770+0000 mon.a (mon.0) 730 : audit [INF] from='osd.7 ' entity='osd.7' cmd={"prefix": "osd crush create-or-move", "id": 7, "weight":0.0195, "args": ["host=vm07", "root=default"]} : dispatch 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: audit 2026-03-21T06:55:49.415312+0000 mon.c (mon.1) 26 : audit [INF] from='client.? 192.168.123.107:0/3612281354' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: audit 2026-03-21T06:55:49.415312+0000 mon.c (mon.1) 26 : audit [INF] from='client.? 192.168.123.107:0/3612281354' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: audit 2026-03-21T06:55:49.417198+0000 mon.a (mon.0) 731 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: audit 2026-03-21T06:55:49.417198+0000 mon.a (mon.0) 731 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"} : dispatch 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: cluster 2026-03-21T06:55:49.603151+0000 mgr.x (mgr.14152) 335 : cluster [DBG] pgmap v275: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:49 vm04 bash[20194]: cluster 2026-03-21T06:55:49.603151+0000 mgr.x (mgr.14152) 335 : cluster [DBG] pgmap v275: 1 pgs: 1 active+clean; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: cluster 2026-03-21T06:55:49.834303+0000 mon.a (mon.0) 732 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: cluster 2026-03-21T06:55:49.834303+0000 mon.a (mon.0) 732 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: cluster 2026-03-21T06:55:49.834325+0000 mon.a (mon.0) 733 : cluster [INF] Cluster is now healthy 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: cluster 2026-03-21T06:55:49.834325+0000 mon.a (mon.0) 733 : cluster [INF] Cluster is now healthy 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: audit 2026-03-21T06:55:49.838079+0000 mon.a (mon.0) 734 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: audit 2026-03-21T06:55:49.838079+0000 mon.a (mon.0) 734 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: cluster 2026-03-21T06:55:49.842810+0000 mon.a (mon.0) 735 : cluster [INF] osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136] boot 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: cluster 2026-03-21T06:55:49.842810+0000 mon.a (mon.0) 735 : cluster [INF] osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136] boot 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: cluster 2026-03-21T06:55:49.842935+0000 mon.a (mon.0) 736 : cluster [DBG] osdmap e56: 8 total, 8 up, 8 in 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: cluster 2026-03-21T06:55:49.842935+0000 mon.a (mon.0) 736 : cluster [DBG] osdmap e56: 8 total, 8 up, 8 in 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: audit 2026-03-21T06:55:49.844003+0000 mon.a (mon.0) 737 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: audit 2026-03-21T06:55:49.844003+0000 mon.a (mon.0) 737 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: audit 2026-03-21T06:55:50.339694+0000 mon.c (mon.1) 27 : audit [INF] from='client.? 192.168.123.107:0/2357448825' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: audit 2026-03-21T06:55:50.339694+0000 mon.c (mon.1) 27 : audit [INF] from='client.? 192.168.123.107:0/2357448825' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: audit 2026-03-21T06:55:50.341381+0000 mon.a (mon.0) 738 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: audit 2026-03-21T06:55:50.341381+0000 mon.a (mon.0) 738 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: audit 2026-03-21T06:55:50.840765+0000 mon.a (mon.0) 739 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-21T06:55:51.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: audit 2026-03-21T06:55:50.840765+0000 mon.a (mon.0) 739 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-21T06:55:51.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: cluster 2026-03-21T06:55:50.843339+0000 mon.a (mon.0) 740 : cluster [DBG] osdmap e57: 8 total, 8 up, 8 in 2026-03-21T06:55:51.140 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:50 vm02 bash[17657]: cluster 2026-03-21T06:55:50.843339+0000 mon.a (mon.0) 740 : cluster [DBG] osdmap e57: 8 total, 8 up, 8 in 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: cluster 2026-03-21T06:55:49.834303+0000 mon.a (mon.0) 732 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: cluster 2026-03-21T06:55:49.834303+0000 mon.a (mon.0) 732 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: cluster 2026-03-21T06:55:49.834325+0000 mon.a (mon.0) 733 : cluster [INF] Cluster is now healthy 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: cluster 2026-03-21T06:55:49.834325+0000 mon.a (mon.0) 733 : cluster [INF] Cluster is now healthy 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: audit 2026-03-21T06:55:49.838079+0000 mon.a (mon.0) 734 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: audit 2026-03-21T06:55:49.838079+0000 mon.a (mon.0) 734 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: cluster 2026-03-21T06:55:49.842810+0000 mon.a (mon.0) 735 : cluster [INF] osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136] boot 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: cluster 2026-03-21T06:55:49.842810+0000 mon.a (mon.0) 735 : cluster [INF] osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136] boot 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: cluster 2026-03-21T06:55:49.842935+0000 mon.a (mon.0) 736 : cluster [DBG] osdmap e56: 8 total, 8 up, 8 in 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: cluster 2026-03-21T06:55:49.842935+0000 mon.a (mon.0) 736 : cluster [DBG] osdmap e56: 8 total, 8 up, 8 in 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: audit 2026-03-21T06:55:49.844003+0000 mon.a (mon.0) 737 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: audit 2026-03-21T06:55:49.844003+0000 mon.a (mon.0) 737 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: audit 2026-03-21T06:55:50.339694+0000 mon.c (mon.1) 27 : audit [INF] from='client.? 192.168.123.107:0/2357448825' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: audit 2026-03-21T06:55:50.339694+0000 mon.c (mon.1) 27 : audit [INF] from='client.? 192.168.123.107:0/2357448825' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: audit 2026-03-21T06:55:50.341381+0000 mon.a (mon.0) 738 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: audit 2026-03-21T06:55:50.341381+0000 mon.a (mon.0) 738 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: audit 2026-03-21T06:55:50.840765+0000 mon.a (mon.0) 739 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: audit 2026-03-21T06:55:50.840765+0000 mon.a (mon.0) 739 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: cluster 2026-03-21T06:55:50.843339+0000 mon.a (mon.0) 740 : cluster [DBG] osdmap e57: 8 total, 8 up, 8 in 2026-03-21T06:55:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:50 vm07 bash[19945]: cluster 2026-03-21T06:55:50.843339+0000 mon.a (mon.0) 740 : cluster [DBG] osdmap e57: 8 total, 8 up, 8 in 2026-03-21T06:55:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: cluster 2026-03-21T06:55:49.834303+0000 mon.a (mon.0) 732 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:55:51.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: cluster 2026-03-21T06:55:49.834303+0000 mon.a (mon.0) 732 : cluster [INF] Health check cleared: OSD_DOWN (was: 1 osds down) 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: cluster 2026-03-21T06:55:49.834325+0000 mon.a (mon.0) 733 : cluster [INF] Cluster is now healthy 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: cluster 2026-03-21T06:55:49.834325+0000 mon.a (mon.0) 733 : cluster [INF] Cluster is now healthy 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: audit 2026-03-21T06:55:49.838079+0000 mon.a (mon.0) 734 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: audit 2026-03-21T06:55:49.838079+0000 mon.a (mon.0) 734 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool create", "pool": "datapool", "pg_num": 3, "pgp_num": 3, "pool_type": "replicated"}]': finished 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: cluster 2026-03-21T06:55:49.842810+0000 mon.a (mon.0) 735 : cluster [INF] osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136] boot 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: cluster 2026-03-21T06:55:49.842810+0000 mon.a (mon.0) 735 : cluster [INF] osd.7 [v2:192.168.123.107:6816/199307136,v1:192.168.123.107:6817/199307136] boot 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: cluster 2026-03-21T06:55:49.842935+0000 mon.a (mon.0) 736 : cluster [DBG] osdmap e56: 8 total, 8 up, 8 in 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: cluster 2026-03-21T06:55:49.842935+0000 mon.a (mon.0) 736 : cluster [DBG] osdmap e56: 8 total, 8 up, 8 in 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: audit 2026-03-21T06:55:49.844003+0000 mon.a (mon.0) 737 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: audit 2026-03-21T06:55:49.844003+0000 mon.a (mon.0) 737 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd metadata", "id": 7} : dispatch 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: audit 2026-03-21T06:55:50.339694+0000 mon.c (mon.1) 27 : audit [INF] from='client.? 192.168.123.107:0/2357448825' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: audit 2026-03-21T06:55:50.339694+0000 mon.c (mon.1) 27 : audit [INF] from='client.? 192.168.123.107:0/2357448825' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: audit 2026-03-21T06:55:50.341381+0000 mon.a (mon.0) 738 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: audit 2026-03-21T06:55:50.341381+0000 mon.a (mon.0) 738 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "osd pool application enable","pool": "datapool","app": "rbd"} : dispatch 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: audit 2026-03-21T06:55:50.840765+0000 mon.a (mon.0) 739 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: audit 2026-03-21T06:55:50.840765+0000 mon.a (mon.0) 739 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "osd pool application enable","pool": "datapool","app": "rbd"}]': finished 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: cluster 2026-03-21T06:55:50.843339+0000 mon.a (mon.0) 740 : cluster [DBG] osdmap e57: 8 total, 8 up, 8 in 2026-03-21T06:55:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:50 vm04 bash[20194]: cluster 2026-03-21T06:55:50.843339+0000 mon.a (mon.0) 740 : cluster [DBG] osdmap e57: 8 total, 8 up, 8 in 2026-03-21T06:55:52.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:51 vm02 bash[17657]: cluster 2026-03-21T06:55:51.603416+0000 mgr.x (mgr.14152) 336 : cluster [DBG] pgmap v278: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:52.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:51 vm02 bash[17657]: cluster 2026-03-21T06:55:51.603416+0000 mgr.x (mgr.14152) 336 : cluster [DBG] pgmap v278: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:52.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:51 vm07 bash[19945]: cluster 2026-03-21T06:55:51.603416+0000 mgr.x (mgr.14152) 336 : cluster [DBG] pgmap v278: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:52.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:51 vm07 bash[19945]: cluster 2026-03-21T06:55:51.603416+0000 mgr.x (mgr.14152) 336 : cluster [DBG] pgmap v278: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:52.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:51 vm04 bash[20194]: cluster 2026-03-21T06:55:51.603416+0000 mgr.x (mgr.14152) 336 : cluster [DBG] pgmap v278: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:52.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:51 vm04 bash[20194]: cluster 2026-03-21T06:55:51.603416+0000 mgr.x (mgr.14152) 336 : cluster [DBG] pgmap v278: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:52.917 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph orch apply iscsi datapool admin admin --trusted_ip_list 192.168.123.102,192.168.123.107 --placement '2;vm02=iscsi.a;vm07=iscsi.b' 2026-03-21T06:55:53.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:52 vm02 bash[17657]: cluster 2026-03-21T06:55:51.857518+0000 mon.a (mon.0) 741 : cluster [DBG] osdmap e58: 8 total, 8 up, 8 in 2026-03-21T06:55:53.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:52 vm02 bash[17657]: cluster 2026-03-21T06:55:51.857518+0000 mon.a (mon.0) 741 : cluster [DBG] osdmap e58: 8 total, 8 up, 8 in 2026-03-21T06:55:53.149 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:55:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:52 vm07 bash[19945]: cluster 2026-03-21T06:55:51.857518+0000 mon.a (mon.0) 741 : cluster [DBG] osdmap e58: 8 total, 8 up, 8 in 2026-03-21T06:55:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:52 vm07 bash[19945]: cluster 2026-03-21T06:55:51.857518+0000 mon.a (mon.0) 741 : cluster [DBG] osdmap e58: 8 total, 8 up, 8 in 2026-03-21T06:55:53.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:52 vm04 bash[20194]: cluster 2026-03-21T06:55:51.857518+0000 mon.a (mon.0) 741 : cluster [DBG] osdmap e58: 8 total, 8 up, 8 in 2026-03-21T06:55:53.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:52 vm04 bash[20194]: cluster 2026-03-21T06:55:51.857518+0000 mon.a (mon.0) 741 : cluster [DBG] osdmap e58: 8 total, 8 up, 8 in 2026-03-21T06:55:53.477 INFO:teuthology.orchestra.run.vm07.stdout:Scheduled iscsi.datapool update... 2026-03-21T06:55:53.530 INFO:tasks.cephadm:Distributing iscsi-gateway.cfg... 2026-03-21T06:55:53.530 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:55:53.530 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/iscsi-gateway.cfg 2026-03-21T06:55:53.537 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:55:53.537 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/iscsi-gateway.cfg 2026-03-21T06:55:53.545 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:55:53.545 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/ceph/iscsi-gateway.cfg 2026-03-21T06:55:53.552 DEBUG:teuthology.orchestra.run.vm02:iscsi.iscsi.a> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@iscsi.iscsi.a.service 2026-03-21T06:55:53.581 DEBUG:teuthology.orchestra.run.vm07:iscsi.iscsi.b> sudo journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@iscsi.iscsi.b.service 2026-03-21T06:55:53.595 INFO:tasks.cephadm:Setting up client nodes... 2026-03-21T06:55:53.595 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph auth get-or-create client.0 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-21T06:55:53.858 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:54.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:53 vm02 bash[17657]: cluster 2026-03-21T06:55:52.865648+0000 mon.a (mon.0) 742 : cluster [DBG] osdmap e59: 8 total, 8 up, 8 in 2026-03-21T06:55:54.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:53 vm02 bash[17657]: cluster 2026-03-21T06:55:52.865648+0000 mon.a (mon.0) 742 : cluster [DBG] osdmap e59: 8 total, 8 up, 8 in 2026-03-21T06:55:54.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:53 vm02 bash[17657]: audit 2026-03-21T06:55:53.471812+0000 mgr.x (mgr.14152) 337 : audit [DBG] from='client.24367 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.102,192.168.123.107", "placement": "2;vm02=iscsi.a;vm07=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:55:54.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:53 vm02 bash[17657]: audit 2026-03-21T06:55:53.471812+0000 mgr.x (mgr.14152) 337 : audit [DBG] from='client.24367 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.102,192.168.123.107", "placement": "2;vm02=iscsi.a;vm07=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:55:54.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:53 vm02 bash[17657]: cephadm 2026-03-21T06:55:53.473025+0000 mgr.x (mgr.14152) 338 : cephadm [INF] Saving service iscsi.datapool spec with placement vm02=iscsi.a;vm07=iscsi.b;count:2 2026-03-21T06:55:54.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:53 vm02 bash[17657]: cephadm 2026-03-21T06:55:53.473025+0000 mgr.x (mgr.14152) 338 : cephadm [INF] Saving service iscsi.datapool spec with placement vm02=iscsi.a;vm07=iscsi.b;count:2 2026-03-21T06:55:54.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:53 vm02 bash[17657]: audit 2026-03-21T06:55:53.478528+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:54.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:53 vm02 bash[17657]: audit 2026-03-21T06:55:53.478528+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:54.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:53 vm02 bash[17657]: cluster 2026-03-21T06:55:53.603664+0000 mgr.x (mgr.14152) 339 : cluster [DBG] pgmap v281: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:54.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:53 vm02 bash[17657]: cluster 2026-03-21T06:55:53.603664+0000 mgr.x (mgr.14152) 339 : cluster [DBG] pgmap v281: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:53 vm07 bash[19945]: cluster 2026-03-21T06:55:52.865648+0000 mon.a (mon.0) 742 : cluster [DBG] osdmap e59: 8 total, 8 up, 8 in 2026-03-21T06:55:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:53 vm07 bash[19945]: cluster 2026-03-21T06:55:52.865648+0000 mon.a (mon.0) 742 : cluster [DBG] osdmap e59: 8 total, 8 up, 8 in 2026-03-21T06:55:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:53 vm07 bash[19945]: audit 2026-03-21T06:55:53.471812+0000 mgr.x (mgr.14152) 337 : audit [DBG] from='client.24367 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.102,192.168.123.107", "placement": "2;vm02=iscsi.a;vm07=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:55:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:53 vm07 bash[19945]: audit 2026-03-21T06:55:53.471812+0000 mgr.x (mgr.14152) 337 : audit [DBG] from='client.24367 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.102,192.168.123.107", "placement": "2;vm02=iscsi.a;vm07=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:55:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:53 vm07 bash[19945]: cephadm 2026-03-21T06:55:53.473025+0000 mgr.x (mgr.14152) 338 : cephadm [INF] Saving service iscsi.datapool spec with placement vm02=iscsi.a;vm07=iscsi.b;count:2 2026-03-21T06:55:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:53 vm07 bash[19945]: cephadm 2026-03-21T06:55:53.473025+0000 mgr.x (mgr.14152) 338 : cephadm [INF] Saving service iscsi.datapool spec with placement vm02=iscsi.a;vm07=iscsi.b;count:2 2026-03-21T06:55:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:53 vm07 bash[19945]: audit 2026-03-21T06:55:53.478528+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:53 vm07 bash[19945]: audit 2026-03-21T06:55:53.478528+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:53 vm07 bash[19945]: cluster 2026-03-21T06:55:53.603664+0000 mgr.x (mgr.14152) 339 : cluster [DBG] pgmap v281: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:53 vm07 bash[19945]: cluster 2026-03-21T06:55:53.603664+0000 mgr.x (mgr.14152) 339 : cluster [DBG] pgmap v281: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:54.234 INFO:teuthology.orchestra.run.vm02.stdout:[client.0] 2026-03-21T06:55:54.235 INFO:teuthology.orchestra.run.vm02.stdout: key = AQD6QL5p7mHwDRAA7Ij0e7wGqrJ3zRDpG5UVhw== 2026-03-21T06:55:54.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:53 vm04 bash[20194]: cluster 2026-03-21T06:55:52.865648+0000 mon.a (mon.0) 742 : cluster [DBG] osdmap e59: 8 total, 8 up, 8 in 2026-03-21T06:55:54.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:53 vm04 bash[20194]: cluster 2026-03-21T06:55:52.865648+0000 mon.a (mon.0) 742 : cluster [DBG] osdmap e59: 8 total, 8 up, 8 in 2026-03-21T06:55:54.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:53 vm04 bash[20194]: audit 2026-03-21T06:55:53.471812+0000 mgr.x (mgr.14152) 337 : audit [DBG] from='client.24367 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.102,192.168.123.107", "placement": "2;vm02=iscsi.a;vm07=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:55:54.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:53 vm04 bash[20194]: audit 2026-03-21T06:55:53.471812+0000 mgr.x (mgr.14152) 337 : audit [DBG] from='client.24367 -' entity='client.admin' cmd=[{"prefix": "orch apply iscsi", "pool": "datapool", "api_user": "admin", "api_password": "admin", "trusted_ip_list": "192.168.123.102,192.168.123.107", "placement": "2;vm02=iscsi.a;vm07=iscsi.b", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:55:54.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:53 vm04 bash[20194]: cephadm 2026-03-21T06:55:53.473025+0000 mgr.x (mgr.14152) 338 : cephadm [INF] Saving service iscsi.datapool spec with placement vm02=iscsi.a;vm07=iscsi.b;count:2 2026-03-21T06:55:54.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:53 vm04 bash[20194]: cephadm 2026-03-21T06:55:53.473025+0000 mgr.x (mgr.14152) 338 : cephadm [INF] Saving service iscsi.datapool spec with placement vm02=iscsi.a;vm07=iscsi.b;count:2 2026-03-21T06:55:54.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:53 vm04 bash[20194]: audit 2026-03-21T06:55:53.478528+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:54.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:53 vm04 bash[20194]: audit 2026-03-21T06:55:53.478528+0000 mon.a (mon.0) 743 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:54.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:53 vm04 bash[20194]: cluster 2026-03-21T06:55:53.603664+0000 mgr.x (mgr.14152) 339 : cluster [DBG] pgmap v281: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:54.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:53 vm04 bash[20194]: cluster 2026-03-21T06:55:53.603664+0000 mgr.x (mgr.14152) 339 : cluster [DBG] pgmap v281: 4 pgs: 2 creating+peering, 1 active+clean, 1 unknown; 449 KiB data, 217 MiB used, 160 GiB / 160 GiB avail 2026-03-21T06:55:54.288 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:55:54.289 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/ceph/ceph.client.0.keyring 2026-03-21T06:55:54.289 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod 0644 /etc/ceph/ceph.client.0.keyring 2026-03-21T06:55:54.300 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph auth get-or-create client.1 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-21T06:55:54.555 INFO:teuthology.orchestra.run.vm04.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.b/config 2026-03-21T06:55:54.925 INFO:teuthology.orchestra.run.vm04.stdout:[client.1] 2026-03-21T06:55:54.925 INFO:teuthology.orchestra.run.vm04.stdout: key = AQD6QL5p2nERNxAAC9PX6ujaO5cnr8AXpEeI7Q== 2026-03-21T06:55:54.980 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:55:54.980 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/ceph/ceph.client.1.keyring 2026-03-21T06:55:54.980 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod 0644 /etc/ceph/ceph.client.1.keyring 2026-03-21T06:55:54.992 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph auth get-or-create client.2 mon 'allow *' osd 'allow *' mds 'allow *' mgr 'allow *' 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.087165+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.087165+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.091507+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.091507+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.233615+0000 mon.a (mon.0) 746 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.233615+0000 mon.a (mon.0) 746 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.236445+0000 mon.a (mon.0) 747 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.236445+0000 mon.a (mon.0) 747 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.923647+0000 mon.a (mon.0) 748 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.923647+0000 mon.a (mon.0) 748 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.926781+0000 mon.a (mon.0) 749 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 bash[17657]: audit 2026-03-21T06:55:54.926781+0000 mon.a (mon.0) 749 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.230 INFO:teuthology.orchestra.run.vm07.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.c/config 2026-03-21T06:55:55.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.087165+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.087165+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.091507+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.091507+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.233615+0000 mon.a (mon.0) 746 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.233615+0000 mon.a (mon.0) 746 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.236445+0000 mon.a (mon.0) 747 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.236445+0000 mon.a (mon.0) 747 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.923647+0000 mon.a (mon.0) 748 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.923647+0000 mon.a (mon.0) 748 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.926781+0000 mon.a (mon.0) 749 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:55 vm04 bash[20194]: audit 2026-03-21T06:55:54.926781+0000 mon.a (mon.0) 749 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.087165+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.087165+0000 mon.a (mon.0) 744 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.091507+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.091507+0000 mon.a (mon.0) 745 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.233615+0000 mon.a (mon.0) 746 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.233615+0000 mon.a (mon.0) 746 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.236445+0000 mon.a (mon.0) 747 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.236445+0000 mon.a (mon.0) 747 : audit [INF] from='client.? 192.168.123.102:0/252747343' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.0", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.923647+0000 mon.a (mon.0) 748 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.923647+0000 mon.a (mon.0) 748 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.926781+0000 mon.a (mon.0) 749 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.281 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:55 vm07 bash[19945]: audit 2026-03-21T06:55:54.926781+0000 mon.a (mon.0) 749 : audit [INF] from='client.? 192.168.123.104:0/2208184041' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.1", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:55.648 INFO:teuthology.orchestra.run.vm07.stdout:[client.2] 2026-03-21T06:55:55.648 INFO:teuthology.orchestra.run.vm07.stdout: key = AQD7QL5prUVuJhAAEiBCNIySc9lrHLtAFTEALw== 2026-03-21T06:55:55.705 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:55:55.705 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/ceph/ceph.client.2.keyring 2026-03-21T06:55:55.705 DEBUG:teuthology.orchestra.run.vm07:> sudo chmod 0644 /etc/ceph/ceph.client.2.keyring 2026-03-21T06:55:55.716 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph config log 1 --format=json 2026-03-21T06:55:55.991 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:56.258 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:55 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:56.258 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:56.258 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:55:55 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:56.258 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:55:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:56.259 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:55:55 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:56.259 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:55:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:56.259 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:55:55 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:56.259 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:55:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:56.495 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:55:56.517 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.433229+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.517 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.433229+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.517 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.437470+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.517 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.437470+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.517 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.442429+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.517 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.442429+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.517 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.446739+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.517 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.446739+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.517 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.447548+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.447548+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.448084+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.448084+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.448547+0000 mon.a (mon.0) 756 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.448547+0000 mon.a (mon.0) 756 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: cephadm 2026-03-21T06:55:55.448918+0000 mgr.x (mgr.14152) 340 : cephadm [INF] Adjusting osd_memory_target on vm07 to 1978M 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: cephadm 2026-03-21T06:55:55.448918+0000 mgr.x (mgr.14152) 340 : cephadm [INF] Adjusting osd_memory_target on vm07 to 1978M 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.454058+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.454058+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.455385+0000 mon.a (mon.0) 758 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.455385+0000 mon.a (mon.0) 758 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.455911+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.455911+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.459407+0000 mon.a (mon.0) 760 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.459407+0000 mon.a (mon.0) 760 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.460431+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.460431+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.462815+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.462815+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.465091+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.465091+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.469871+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.469871+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: cephadm 2026-03-21T06:55:55.470456+0000 mgr.x (mgr.14152) 341 : cephadm [INF] Deploying daemon iscsi.iscsi.a on vm02 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: cephadm 2026-03-21T06:55:55.470456+0000 mgr.x (mgr.14152) 341 : cephadm [INF] Deploying daemon iscsi.iscsi.a on vm02 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: cluster 2026-03-21T06:55:55.603973+0000 mgr.x (mgr.14152) 342 : cluster [DBG] pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 355 B/s wr, 0 op/s 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: cluster 2026-03-21T06:55:55.603973+0000 mgr.x (mgr.14152) 342 : cluster [DBG] pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 355 B/s wr, 0 op/s 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.642605+0000 mon.c (mon.1) 28 : audit [INF] from='client.? 192.168.123.107:0/1019988500' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.642605+0000 mon.c (mon.1) 28 : audit [INF] from='client.? 192.168.123.107:0/1019988500' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.644436+0000 mon.a (mon.0) 765 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.644436+0000 mon.a (mon.0) 765 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.647455+0000 mon.a (mon.0) 766 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:55.647455+0000 mon.a (mon.0) 766 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.294982+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.294982+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.299785+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.299785+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.307686+0000 mon.a (mon.0) 769 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.307686+0000 mon.a (mon.0) 769 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.308302+0000 mon.a (mon.0) 770 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.308302+0000 mon.a (mon.0) 770 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.312818+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.312818+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.316961+0000 mon.a (mon.0) 772 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.518 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[17657]: audit 2026-03-21T06:55:56.316961+0000 mon.a (mon.0) 772 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.570 INFO:teuthology.orchestra.run.vm02.stdout:[{"version":15,"timestamp":"2026-03-21T06:55:55.449106+0000","name":"","changes":[{"name":"osd/host:vm07/osd_memory_target","previous_value":"6223219097","new_value":"2074406365"}]}] 2026-03-21T06:55:56.570 INFO:tasks.ceph_manager:config epoch is 15 2026-03-21T06:55:56.570 INFO:tasks.ceph:Waiting until ceph daemons up and pgs clean... 2026-03-21T06:55:56.570 INFO:tasks.cephadm.ceph_manager.ceph:waiting for mgr available 2026-03-21T06:55:56.570 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph mgr dump --format=json 2026-03-21T06:55:56.746 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.433229+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.746 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.433229+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.746 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.437470+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.746 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.437470+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.746 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.442429+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.746 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.442429+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.746 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.446739+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.446739+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.447548+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.447548+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.448084+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.448084+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.448547+0000 mon.a (mon.0) 756 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.448547+0000 mon.a (mon.0) 756 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: cephadm 2026-03-21T06:55:55.448918+0000 mgr.x (mgr.14152) 340 : cephadm [INF] Adjusting osd_memory_target on vm07 to 1978M 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: cephadm 2026-03-21T06:55:55.448918+0000 mgr.x (mgr.14152) 340 : cephadm [INF] Adjusting osd_memory_target on vm07 to 1978M 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.454058+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.454058+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.455385+0000 mon.a (mon.0) 758 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.455385+0000 mon.a (mon.0) 758 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.455911+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.455911+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.459407+0000 mon.a (mon.0) 760 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.459407+0000 mon.a (mon.0) 760 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.460431+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.460431+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.462815+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.462815+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.465091+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.465091+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.469871+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.469871+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: cephadm 2026-03-21T06:55:55.470456+0000 mgr.x (mgr.14152) 341 : cephadm [INF] Deploying daemon iscsi.iscsi.a on vm02 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: cephadm 2026-03-21T06:55:55.470456+0000 mgr.x (mgr.14152) 341 : cephadm [INF] Deploying daemon iscsi.iscsi.a on vm02 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: cluster 2026-03-21T06:55:55.603973+0000 mgr.x (mgr.14152) 342 : cluster [DBG] pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 355 B/s wr, 0 op/s 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: cluster 2026-03-21T06:55:55.603973+0000 mgr.x (mgr.14152) 342 : cluster [DBG] pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 355 B/s wr, 0 op/s 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.642605+0000 mon.c (mon.1) 28 : audit [INF] from='client.? 192.168.123.107:0/1019988500' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.642605+0000 mon.c (mon.1) 28 : audit [INF] from='client.? 192.168.123.107:0/1019988500' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.644436+0000 mon.a (mon.0) 765 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.644436+0000 mon.a (mon.0) 765 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.647455+0000 mon.a (mon.0) 766 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:55.647455+0000 mon.a (mon.0) 766 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.294982+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.294982+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.299785+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.299785+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.307686+0000 mon.a (mon.0) 769 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.307686+0000 mon.a (mon.0) 769 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.308302+0000 mon.a (mon.0) 770 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.308302+0000 mon.a (mon.0) 770 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.312818+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.312818+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.316961+0000 mon.a (mon.0) 772 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.747 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 bash[19945]: audit 2026-03-21T06:55:56.316961+0000 mon.a (mon.0) 772 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.433229+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.433229+0000 mon.a (mon.0) 750 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.437470+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.437470+0000 mon.a (mon.0) 751 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.442429+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.442429+0000 mon.a (mon.0) 752 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.446739+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.446739+0000 mon.a (mon.0) 753 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.447548+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.447548+0000 mon.a (mon.0) 754 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.5", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.448084+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.448084+0000 mon.a (mon.0) 755 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.7", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.448547+0000 mon.a (mon.0) 756 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.448547+0000 mon.a (mon.0) 756 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.6", "name": "osd_memory_target"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: cephadm 2026-03-21T06:55:55.448918+0000 mgr.x (mgr.14152) 340 : cephadm [INF] Adjusting osd_memory_target on vm07 to 1978M 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: cephadm 2026-03-21T06:55:55.448918+0000 mgr.x (mgr.14152) 340 : cephadm [INF] Adjusting osd_memory_target on vm07 to 1978M 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.454058+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.454058+0000 mon.a (mon.0) 757 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.455385+0000 mon.a (mon.0) 758 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.455385+0000 mon.a (mon.0) 758 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.455911+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.455911+0000 mon.a (mon.0) 759 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.459407+0000 mon.a (mon.0) 760 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.459407+0000 mon.a (mon.0) 760 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.460431+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.460431+0000 mon.a (mon.0) 761 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.462815+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.462815+0000 mon.a (mon.0) 762 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.465091+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.465091+0000 mon.a (mon.0) 763 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.a", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.469871+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.469871+0000 mon.a (mon.0) 764 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: cephadm 2026-03-21T06:55:55.470456+0000 mgr.x (mgr.14152) 341 : cephadm [INF] Deploying daemon iscsi.iscsi.a on vm02 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: cephadm 2026-03-21T06:55:55.470456+0000 mgr.x (mgr.14152) 341 : cephadm [INF] Deploying daemon iscsi.iscsi.a on vm02 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: cluster 2026-03-21T06:55:55.603973+0000 mgr.x (mgr.14152) 342 : cluster [DBG] pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 355 B/s wr, 0 op/s 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: cluster 2026-03-21T06:55:55.603973+0000 mgr.x (mgr.14152) 342 : cluster [DBG] pgmap v282: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 355 B/s wr, 0 op/s 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.642605+0000 mon.c (mon.1) 28 : audit [INF] from='client.? 192.168.123.107:0/1019988500' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.642605+0000 mon.c (mon.1) 28 : audit [INF] from='client.? 192.168.123.107:0/1019988500' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.644436+0000 mon.a (mon.0) 765 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.644436+0000 mon.a (mon.0) 765 : audit [INF] from='client.? ' entity='client.admin' cmd={"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]} : dispatch 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.647455+0000 mon.a (mon.0) 766 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:55.647455+0000 mon.a (mon.0) 766 : audit [INF] from='client.? ' entity='client.admin' cmd='[{"prefix": "auth get-or-create", "entity": "client.2", "caps": ["mon", "allow *", "osd", "allow *", "mds", "allow *", "mgr", "allow *"]}]': finished 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.294982+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.294982+0000 mon.a (mon.0) 767 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.299785+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.299785+0000 mon.a (mon.0) 768 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.307686+0000 mon.a (mon.0) 769 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.307686+0000 mon.a (mon.0) 769 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:56.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.308302+0000 mon.a (mon.0) 770 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.308302+0000 mon.a (mon.0) 770 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]} : dispatch 2026-03-21T06:55:56.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.312818+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.312818+0000 mon.a (mon.0) 771 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd='[{"prefix": "auth get-or-create", "entity": "client.iscsi.iscsi.b", "caps": ["mon", "profile rbd, allow command \"osd blocklist\", allow command \"config-key get\" with \"key\" prefix \"iscsi/\"", "mgr", "allow command \"service status\"", "osd", "allow rwx"]}]': finished 2026-03-21T06:55:56.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.316961+0000 mon.a (mon.0) 772 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:56 vm04 bash[20194]: audit 2026-03-21T06:55:56.316961+0000 mon.a (mon.0) 772 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:55:56.823 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[49158]: debug Processing osd blocklist entries for this node 2026-03-21T06:55:56.823 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:55:56 vm02 bash[49158]: debug Removing blocklisted entry for this host : 192.168.123.102:6801/4266123507 2026-03-21T06:55:56.828 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:57.058 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:56 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:57.058 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:55:56 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:57.058 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:56 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:57.059 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:55:56 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:57.242 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:55:57.327 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":14,"flags":0,"active_gid":14152,"active_name":"x","active_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6800","nonce":512722970},{"type":"v1","addr":"192.168.123.102:6801","nonce":512722970}]},"active_addr":"192.168.123.102:6801/512722970","active_change":"2026-03-21T06:48:09.529048+0000","active_mgr_features":4544132024016699391,"available":true,"standbys":[],"modules":["cephadm","dashboard","iostat","nfs"],"available_modules":[{"name":"alerts","can_run":true,"error_string":"","module_options":{"interval":{"name":"interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"How frequently to reexamine health status","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"smtp_destination":{"name":"smtp_destination","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Email address to send alerts to, use commas to separate multiple","long_desc":"","tags":[],"see_also":[]},"smtp_from_name":{"name":"smtp_from_name","type":"str","level":"advanced","flags":1,"default_value":"Ceph","min":"","max":"","enum_allowed":[],"desc":"Email From: name","long_desc":"","tags":[],"see_also":[]},"smtp_host":{"name":"smtp_host","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_password":{"name":"smtp_password","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Password to authenticate with","long_desc":"","tags":[],"see_also":[]},"smtp_port":{"name":"smtp_port","type":"int","level":"advanced","flags":1,"default_value":"465","min":"","max":"","enum_allowed":[],"desc":"SMTP port","long_desc":"","tags":[],"see_also":[]},"smtp_sender":{"name":"smtp_sender","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"SMTP envelope sender","long_desc":"","tags":[],"see_also":[]},"smtp_ssl":{"name":"smtp_ssl","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Use SSL to connect to SMTP server","long_desc":"","tags":[],"see_also":[]},"smtp_user":{"name":"smtp_user","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"User to authenticate as","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"balancer","can_run":true,"error_string":"","module_options":{"active":{"name":"active","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically balance PGs across cluster","long_desc":"","tags":[],"see_also":[]},"begin_time":{"name":"begin_time","type":"str","level":"advanced","flags":1,"default_value":"0000","min":"","max":"","enum_allowed":[],"desc":"beginning time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"begin_weekday":{"name":"begin_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to this day of the week or later","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"crush_compat_max_iterations":{"name":"crush_compat_max_iterations","type":"uint","level":"advanced","flags":1,"default_value":"25","min":"1","max":"250","enum_allowed":[],"desc":"maximum number of iterations to attempt optimization","long_desc":"","tags":[],"see_also":[]},"crush_compat_metrics":{"name":"crush_compat_metrics","type":"str","level":"advanced","flags":1,"default_value":"pgs,objects,bytes","min":"","max":"","enum_allowed":[],"desc":"metrics with which to calculate OSD utilization","long_desc":"Value is a list of one or more of \"pgs\", \"objects\", or \"bytes\", and indicates which metrics to use to balance utilization.","tags":[],"see_also":[]},"crush_compat_step":{"name":"crush_compat_step","type":"float","level":"advanced","flags":1,"default_value":"0.5","min":"0.001","max":"0.999","enum_allowed":[],"desc":"aggressiveness of optimization","long_desc":".99 is very aggressive, .01 is less aggressive","tags":[],"see_also":[]},"end_time":{"name":"end_time","type":"str","level":"advanced","flags":1,"default_value":"2359","min":"","max":"","enum_allowed":[],"desc":"ending time of day to automatically balance","long_desc":"This is a time of day in the format HHMM.","tags":[],"see_also":[]},"end_weekday":{"name":"end_weekday","type":"uint","level":"advanced","flags":1,"default_value":"0","min":"0","max":"6","enum_allowed":[],"desc":"Restrict automatic balancing to days of the week earlier than this","long_desc":"0 = Sunday, 1 = Monday, etc.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_score":{"name":"min_score","type":"float","level":"advanced","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"minimum score, below which no optimization is attempted","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":1,"default_value":"upmap","min":"","max":"","enum_allowed":["crush-compat","none","read","upmap","upmap-read"],"desc":"Balancer mode","long_desc":"","tags":[],"see_also":[]},"pool_ids":{"name":"pool_ids","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"pools which the automatic balancing will be limited to","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and attempt optimization","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_pg_upmap_activity":{"name":"update_pg_upmap_activity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Updates pg_upmap activity stats to be used in `balancer status detail`","long_desc":"","tags":[],"see_also":[]},"upmap_max_deviation":{"name":"upmap_max_deviation","type":"int","level":"advanced","flags":1,"default_value":"5","min":"1","max":"","enum_allowed":[],"desc":"deviation below which no optimization is attempted","long_desc":"If the number of PGs are within this count then no optimization is attempted","tags":[],"see_also":[]},"upmap_max_optimizations":{"name":"upmap_max_optimizations","type":"uint","level":"advanced","flags":1,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"maximum upmap optimizations to make per attempt","long_desc":"","tags":[],"see_also":[]}}},{"name":"cephadm","can_run":true,"error_string":"","module_options":{"agent_down_multiplier":{"name":"agent_down_multiplier","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"","max":"","enum_allowed":[],"desc":"Multiplied by agent refresh rate to calculate how long agent must not report before being marked down","long_desc":"","tags":[],"see_also":[]},"agent_refresh_rate":{"name":"agent_refresh_rate","type":"secs","level":"advanced","flags":0,"default_value":"20","min":"","max":"","enum_allowed":[],"desc":"How often agent on each host will try to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"agent_starting_port":{"name":"agent_starting_port","type":"int","level":"advanced","flags":0,"default_value":"4721","min":"","max":"","enum_allowed":[],"desc":"First port agent will try to bind to (will also try up to next 1000 subsequent ports if blocked)","long_desc":"","tags":[],"see_also":[]},"allow_ptrace":{"name":"allow_ptrace","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow SYS_PTRACE capability on ceph containers","long_desc":"The SYS_PTRACE capability is needed to attach to a process with gdb or strace. Enabling this options can allow debugging daemons that encounter problems at runtime.","tags":[],"see_also":[]},"autotune_interval":{"name":"autotune_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to autotune daemon memory","long_desc":"","tags":[],"see_also":[]},"autotune_memory_target_ratio":{"name":"autotune_memory_target_ratio","type":"float","level":"advanced","flags":0,"default_value":"0.7","min":"","max":"","enum_allowed":[],"desc":"ratio of total system memory to divide amongst autotuned daemons","long_desc":"","tags":[],"see_also":[]},"cephadm_log_destination":{"name":"cephadm_log_destination","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":["file","file,syslog","syslog"],"desc":"Destination for cephadm command's persistent logging","long_desc":"","tags":[],"see_also":[]},"certificate_automated_rotation_enabled":{"name":"certificate_automated_rotation_enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"This flag controls whether cephadm automatically rotates certificates upon expiration.","long_desc":"","tags":[],"see_also":[]},"certificate_check_debug_mode":{"name":"certificate_check_debug_mode","type":"bool","level":"dev","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"FOR TESTING ONLY: This flag forces the certificate check instead of waiting for certificate_check_period.","long_desc":"","tags":[],"see_also":[]},"certificate_check_period":{"name":"certificate_check_period","type":"int","level":"advanced","flags":0,"default_value":"1","min":"0","max":"30","enum_allowed":[],"desc":"Specifies how often (in days) the certificate should be checked for validity.","long_desc":"","tags":[],"see_also":[]},"certificate_duration_days":{"name":"certificate_duration_days","type":"int","level":"advanced","flags":0,"default_value":"1095","min":"90","max":"3650","enum_allowed":[],"desc":"Specifies the duration of self certificates generated and signed by cephadm root CA","long_desc":"","tags":[],"see_also":[]},"certificate_renewal_threshold_days":{"name":"certificate_renewal_threshold_days","type":"int","level":"advanced","flags":0,"default_value":"30","min":"10","max":"90","enum_allowed":[],"desc":"Specifies the lead time in days to initiate certificate renewal before expiration.","long_desc":"","tags":[],"see_also":[]},"cgroups_split":{"name":"cgroups_split","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Pass --cgroups=split when cephadm creates containers (currently podman only)","long_desc":"","tags":[],"see_also":[]},"config_checks_enabled":{"name":"config_checks_enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable or disable the cephadm configuration analysis","long_desc":"","tags":[],"see_also":[]},"config_dashboard":{"name":"config_dashboard","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"manage configs like API endpoints in Dashboard.","long_desc":"","tags":[],"see_also":[]},"container_image_alertmanager":{"name":"container_image_alertmanager","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/alertmanager:v0.28.1","min":"","max":"","enum_allowed":[],"desc":"Alertmanager container image","long_desc":"","tags":[],"see_also":[]},"container_image_base":{"name":"container_image_base","type":"str","level":"advanced","flags":1,"default_value":"quay.io/ceph/ceph","min":"","max":"","enum_allowed":[],"desc":"Container image name, without the tag","long_desc":"","tags":[],"see_also":[]},"container_image_elasticsearch":{"name":"container_image_elasticsearch","type":"str","level":"advanced","flags":0,"default_value":"quay.io/omrizeneva/elasticsearch:6.8.23","min":"","max":"","enum_allowed":[],"desc":"Elasticsearch container image","long_desc":"","tags":[],"see_also":[]},"container_image_grafana":{"name":"container_image_grafana","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/grafana:12.3.1","min":"","max":"","enum_allowed":[],"desc":"Grafana container image","long_desc":"","tags":[],"see_also":[]},"container_image_haproxy":{"name":"container_image_haproxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/haproxy:2.3","min":"","max":"","enum_allowed":[],"desc":"Haproxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_agent":{"name":"container_image_jaeger_agent","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-agent:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger agent container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_collector":{"name":"container_image_jaeger_collector","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-collector:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger collector container image","long_desc":"","tags":[],"see_also":[]},"container_image_jaeger_query":{"name":"container_image_jaeger_query","type":"str","level":"advanced","flags":0,"default_value":"quay.io/jaegertracing/jaeger-query:1.29","min":"","max":"","enum_allowed":[],"desc":"Jaeger query container image","long_desc":"","tags":[],"see_also":[]},"container_image_keepalived":{"name":"container_image_keepalived","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/keepalived:2.2.4","min":"","max":"","enum_allowed":[],"desc":"Keepalived container image","long_desc":"","tags":[],"see_also":[]},"container_image_loki":{"name":"container_image_loki","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/loki:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Loki container image","long_desc":"","tags":[],"see_also":[]},"container_image_nginx":{"name":"container_image_nginx","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nginx:sclorg-nginx-126","min":"","max":"","enum_allowed":[],"desc":"Nginx container image","long_desc":"","tags":[],"see_also":[]},"container_image_node_exporter":{"name":"container_image_node_exporter","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/node-exporter:v1.9.1","min":"","max":"","enum_allowed":[],"desc":"Node exporter container image","long_desc":"","tags":[],"see_also":[]},"container_image_nvmeof":{"name":"container_image_nvmeof","type":"str","level":"advanced","flags":0,"default_value":"quay.io/ceph/nvmeof:1.5","min":"","max":"","enum_allowed":[],"desc":"Nvmeof container image","long_desc":"","tags":[],"see_also":[]},"container_image_oauth2_proxy":{"name":"container_image_oauth2_proxy","type":"str","level":"advanced","flags":0,"default_value":"quay.io/oauth2-proxy/oauth2-proxy:v7.6.0","min":"","max":"","enum_allowed":[],"desc":"Oauth2 proxy container image","long_desc":"","tags":[],"see_also":[]},"container_image_prometheus":{"name":"container_image_prometheus","type":"str","level":"advanced","flags":0,"default_value":"quay.io/prometheus/prometheus:v3.6.0","min":"","max":"","enum_allowed":[],"desc":"Prometheus container image","long_desc":"","tags":[],"see_also":[]},"container_image_promtail":{"name":"container_image_promtail","type":"str","level":"advanced","flags":0,"default_value":"docker.io/grafana/promtail:3.0.0","min":"","max":"","enum_allowed":[],"desc":"Promtail container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba":{"name":"container_image_samba","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-server:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba container image","long_desc":"","tags":[],"see_also":[]},"container_image_samba_metrics":{"name":"container_image_samba_metrics","type":"str","level":"advanced","flags":0,"default_value":"quay.io/samba.org/samba-metrics:ceph20-centos-amd64","min":"","max":"","enum_allowed":[],"desc":"Samba metrics container image","long_desc":"","tags":[],"see_also":[]},"container_image_snmp_gateway":{"name":"container_image_snmp_gateway","type":"str","level":"advanced","flags":0,"default_value":"docker.io/maxwo/snmp-notifier:v1.2.1","min":"","max":"","enum_allowed":[],"desc":"Snmp gateway container image","long_desc":"","tags":[],"see_also":[]},"container_init":{"name":"container_init","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Run podman/docker with `--init`","long_desc":"","tags":[],"see_also":[]},"daemon_cache_timeout":{"name":"daemon_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"seconds to cache service (daemon) inventory","long_desc":"","tags":[],"see_also":[]},"default_cephadm_command_timeout":{"name":"default_cephadm_command_timeout","type":"int","level":"advanced","flags":0,"default_value":"900","min":"","max":"","enum_allowed":[],"desc":"Default timeout applied to cephadm commands run directly on the host (in seconds)","long_desc":"","tags":[],"see_also":[]},"default_registry":{"name":"default_registry","type":"str","level":"advanced","flags":0,"default_value":"quay.io","min":"","max":"","enum_allowed":[],"desc":"Search-registry to which we should normalize unqualified image names. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"device_cache_timeout":{"name":"device_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"seconds to cache device inventory","long_desc":"","tags":[],"see_also":[]},"device_enhanced_scan":{"name":"device_enhanced_scan","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use libstoragemgmt during device scans","long_desc":"","tags":[],"see_also":[]},"facts_cache_timeout":{"name":"facts_cache_timeout","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"seconds to cache host facts data","long_desc":"","tags":[],"see_also":[]},"grafana_dashboards_path":{"name":"grafana_dashboards_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/grafana/dashboards/ceph-dashboard/","min":"","max":"","enum_allowed":[],"desc":"location of dashboards to include in grafana deployments","long_desc":"","tags":[],"see_also":[]},"host_check_interval":{"name":"host_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to perform a host check","long_desc":"","tags":[],"see_also":[]},"hw_monitoring":{"name":"hw_monitoring","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Deploy hw monitoring daemon on every host.","long_desc":"","tags":[],"see_also":[]},"inventory_list_all":{"name":"inventory_list_all","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Whether ceph-volume inventory should report more devices (mostly mappers (LVs / mpaths), partitions...)","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_refresh_metadata":{"name":"log_refresh_metadata","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Log all refresh metadata. Includes daemon, device, and host info collected regularly. Only has effect if logging at debug level","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"log to the \"cephadm\" cluster log channel\"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf":{"name":"manage_etc_ceph_ceph_conf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Manage and own /etc/ceph/ceph.conf on the hosts.","long_desc":"","tags":[],"see_also":[]},"manage_etc_ceph_ceph_conf_hosts":{"name":"manage_etc_ceph_ceph_conf_hosts","type":"str","level":"advanced","flags":0,"default_value":"*","min":"","max":"","enum_allowed":[],"desc":"PlacementSpec describing on which hosts to manage /etc/ceph/ceph.conf","long_desc":"","tags":[],"see_also":[]},"max_count_per_host":{"name":"max_count_per_host","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of daemons per service per host","long_desc":"","tags":[],"see_also":[]},"max_osd_draining_count":{"name":"max_osd_draining_count","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"max number of osds that will be drained simultaneously when osds are removed","long_desc":"","tags":[],"see_also":[]},"migration_current":{"name":"migration_current","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"internal - do not modify","long_desc":"","tags":[],"see_also":[]},"mode":{"name":"mode","type":"str","level":"advanced","flags":0,"default_value":"root","min":"","max":"","enum_allowed":["cephadm-package","root"],"desc":"mode for remote execution of cephadm","long_desc":"","tags":[],"see_also":[]},"oob_default_addr":{"name":"oob_default_addr","type":"str","level":"advanced","flags":0,"default_value":"169.254.1.1","min":"","max":"","enum_allowed":[],"desc":"Default address for RedFish API (oob management).","long_desc":"","tags":[],"see_also":[]},"prometheus_alerts_path":{"name":"prometheus_alerts_path","type":"str","level":"advanced","flags":0,"default_value":"/etc/prometheus/ceph/ceph_default_alerts.yml","min":"","max":"","enum_allowed":[],"desc":"location of alerts to include in prometheus deployments","long_desc":"","tags":[],"see_also":[]},"registry_insecure":{"name":"registry_insecure","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Registry is to be considered insecure (no TLS available). Only for development purposes.","long_desc":"","tags":[],"see_also":[]},"registry_password":{"name":"registry_password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository password. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"registry_url":{"name":"registry_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Registry url for login purposes. This is not the default registry","long_desc":"","tags":[],"see_also":[]},"registry_username":{"name":"registry_username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"Custom repository username. Only used for logging into a registry.","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"service_discovery_port":{"name":"service_discovery_port","type":"int","level":"advanced","flags":0,"default_value":"8765","min":"","max":"","enum_allowed":[],"desc":"cephadm service discovery port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssh_config_file":{"name":"ssh_config_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"customized SSH config file to connect to managed hosts","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_count_max":{"name":"ssh_keepalive_count_max","type":"int","level":"advanced","flags":0,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"How many times ssh connections can fail liveness checks before the host is marked offline","long_desc":"","tags":[],"see_also":[]},"ssh_keepalive_interval":{"name":"ssh_keepalive_interval","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"How often ssh connections are checked for liveness","long_desc":"","tags":[],"see_also":[]},"stray_daemon_check_interval":{"name":"stray_daemon_check_interval","type":"secs","level":"advanced","flags":0,"default_value":"1800","min":"","max":"","enum_allowed":[],"desc":"how frequently cephadm should check for the presence of stray daemons","long_desc":"","tags":[],"see_also":[]},"use_agent":{"name":"use_agent","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Use cephadm agent on each host to gather and send metadata","long_desc":"","tags":[],"see_also":[]},"use_repo_digest":{"name":"use_repo_digest","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Automatically convert image tags to image digest. Make sure all daemons use the same image","long_desc":"","tags":[],"see_also":[]},"warn_on_failed_host_check":{"name":"warn_on_failed_host_check","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if the host check fails","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_daemons":{"name":"warn_on_stray_daemons","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected that are not managed by cephadm","long_desc":"","tags":[],"see_also":[]},"warn_on_stray_hosts":{"name":"warn_on_stray_hosts","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"raise a health warning if daemons are detected on a host that is not managed by cephadm","long_desc":"","tags":[],"see_also":[]}}},{"name":"crash","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"retain_interval":{"name":"retain_interval","type":"secs","level":"advanced","flags":1,"default_value":"31536000","min":"","max":"","enum_allowed":[],"desc":"how long to retain crashes before pruning them","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_recent_interval":{"name":"warn_recent_interval","type":"secs","level":"advanced","flags":1,"default_value":"1209600","min":"","max":"","enum_allowed":[],"desc":"time interval in which to warn about recent crashes","long_desc":"","tags":[],"see_also":[]}}},{"name":"dashboard","can_run":true,"error_string":"","module_options":{"ACCOUNT_LOCKOUT_ATTEMPTS":{"name":"ACCOUNT_LOCKOUT_ATTEMPTS","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_HOST":{"name":"ALERTMANAGER_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ALERTMANAGER_API_SSL_VERIFY":{"name":"ALERTMANAGER_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_ENABLED":{"name":"AUDIT_API_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"AUDIT_API_LOG_PAYLOAD":{"name":"AUDIT_API_LOG_PAYLOAD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ENABLE_BROWSABLE_API":{"name":"ENABLE_BROWSABLE_API","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_CEPHFS":{"name":"FEATURE_TOGGLE_CEPHFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_DASHBOARD":{"name":"FEATURE_TOGGLE_DASHBOARD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_ISCSI":{"name":"FEATURE_TOGGLE_ISCSI","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_MIRRORING":{"name":"FEATURE_TOGGLE_MIRRORING","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_NFS":{"name":"FEATURE_TOGGLE_NFS","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RBD":{"name":"FEATURE_TOGGLE_RBD","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"FEATURE_TOGGLE_RGW":{"name":"FEATURE_TOGGLE_RGW","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE":{"name":"GANESHA_CLUSTERS_RADOS_POOL_NAMESPACE","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_PASSWORD":{"name":"GRAFANA_API_PASSWORD","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_SSL_VERIFY":{"name":"GRAFANA_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_URL":{"name":"GRAFANA_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_API_USERNAME":{"name":"GRAFANA_API_USERNAME","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_FRONTEND_API_URL":{"name":"GRAFANA_FRONTEND_API_URL","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"GRAFANA_UPDATE_DASHBOARDS":{"name":"GRAFANA_UPDATE_DASHBOARDS","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISCSI_API_SSL_VERIFICATION":{"name":"ISCSI_API_SSL_VERIFICATION","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ISSUE_TRACKER_API_KEY":{"name":"ISSUE_TRACKER_API_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MANAGED_BY_CLUSTERS":{"name":"MANAGED_BY_CLUSTERS","type":"str","level":"advanced","flags":0,"default_value":"[]","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"MULTICLUSTER_CONFIG":{"name":"MULTICLUSTER_CONFIG","type":"str","level":"advanced","flags":0,"default_value":"{}","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_HOST":{"name":"PROMETHEUS_API_HOST","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROMETHEUS_API_SSL_VERIFY":{"name":"PROMETHEUS_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PROM_ALERT_CREDENTIAL_CACHE_TTL":{"name":"PROM_ALERT_CREDENTIAL_CACHE_TTL","type":"int","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_COMPLEXITY_ENABLED":{"name":"PWD_POLICY_CHECK_COMPLEXITY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED":{"name":"PWD_POLICY_CHECK_EXCLUSION_LIST_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_LENGTH_ENABLED":{"name":"PWD_POLICY_CHECK_LENGTH_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_OLDPWD_ENABLED":{"name":"PWD_POLICY_CHECK_OLDPWD_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_REPETITIVE_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED":{"name":"PWD_POLICY_CHECK_SEQUENTIAL_CHARS_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_CHECK_USERNAME_ENABLED":{"name":"PWD_POLICY_CHECK_USERNAME_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_ENABLED":{"name":"PWD_POLICY_ENABLED","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_EXCLUSION_LIST":{"name":"PWD_POLICY_EXCLUSION_LIST","type":"str","level":"advanced","flags":0,"default_value":"osd,host,dashboard,pool,block,nfs,ceph,monitors,gateway,logs,crush,maps","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_COMPLEXITY":{"name":"PWD_POLICY_MIN_COMPLEXITY","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"PWD_POLICY_MIN_LENGTH":{"name":"PWD_POLICY_MIN_LENGTH","type":"int","level":"advanced","flags":0,"default_value":"8","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"REST_REQUESTS_TIMEOUT":{"name":"REST_REQUESTS_TIMEOUT","type":"int","level":"advanced","flags":0,"default_value":"45","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ACCESS_KEY":{"name":"RGW_API_ACCESS_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_ADMIN_RESOURCE":{"name":"RGW_API_ADMIN_RESOURCE","type":"str","level":"advanced","flags":0,"default_value":"admin","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SECRET_KEY":{"name":"RGW_API_SECRET_KEY","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_API_SSL_VERIFY":{"name":"RGW_API_SSL_VERIFY","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"RGW_HOSTNAME_PER_DAEMON":{"name":"RGW_HOSTNAME_PER_DAEMON","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"UNSAFE_TLS_v1_2":{"name":"UNSAFE_TLS_v1_2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_SPAN":{"name":"USER_PWD_EXPIRATION_SPAN","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_1":{"name":"USER_PWD_EXPIRATION_WARNING_1","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"USER_PWD_EXPIRATION_WARNING_2":{"name":"USER_PWD_EXPIRATION_WARNING_2","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"cross_origin_url":{"name":"cross_origin_url","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crt_file":{"name":"crt_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"crypto_caller":{"name":"crypto_caller","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"debug":{"name":"debug","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable/disable debug options","long_desc":"","tags":[],"see_also":[]},"jwt_token_ttl":{"name":"jwt_token_ttl","type":"int","level":"advanced","flags":0,"default_value":"28800","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"key_file":{"name":"key_file","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"motd":{"name":"motd","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"The message of the day","long_desc":"","tags":[],"see_also":[]},"redirect_resolve_ip_addr":{"name":"redirect_resolve_ip_addr","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":0,"default_value":"8080","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl_server_port":{"name":"ssl_server_port","type":"int","level":"advanced","flags":0,"default_value":"8443","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sso_oauth2":{"name":"sso_oauth2","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":0,"default_value":"redirect","min":"","max":"","enum_allowed":["error","redirect"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":0,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url_prefix":{"name":"url_prefix","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"devicehealth","can_run":true,"error_string":"","module_options":{"enable_monitoring":{"name":"enable_monitoring","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"monitor device health metrics","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mark_out_threshold":{"name":"mark_out_threshold","type":"secs","level":"advanced","flags":1,"default_value":"2419200","min":"","max":"","enum_allowed":[],"desc":"automatically mark OSD if it may fail before this long","long_desc":"","tags":[],"see_also":[]},"pool_name":{"name":"pool_name","type":"str","level":"advanced","flags":1,"default_value":"device_health_metrics","min":"","max":"","enum_allowed":[],"desc":"name of pool in which to store device health metrics","long_desc":"","tags":[],"see_also":[]},"retention_period":{"name":"retention_period","type":"secs","level":"advanced","flags":1,"default_value":"15552000","min":"","max":"","enum_allowed":[],"desc":"how long to retain device health metrics","long_desc":"","tags":[],"see_also":[]},"scrape_frequency":{"name":"scrape_frequency","type":"secs","level":"advanced","flags":1,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"how frequently to scrape device health metrics","long_desc":"","tags":[],"see_also":[]},"self_heal":{"name":"self_heal","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"preemptively heal cluster around devices that may fail","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"how frequently to wake up and check device health","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"warn_threshold":{"name":"warn_threshold","type":"secs","level":"advanced","flags":1,"default_value":"7257600","min":"","max":"","enum_allowed":[],"desc":"raise health warning if OSD may fail before this long","long_desc":"","tags":[],"see_also":[]}}},{"name":"diskprediction_local","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predict_interval":{"name":"predict_interval","type":"str","level":"advanced","flags":0,"default_value":"86400","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"predictor_model":{"name":"predictor_model","type":"str","level":"advanced","flags":0,"default_value":"prophetstor","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"str","level":"advanced","flags":0,"default_value":"600","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"influx","can_run":false,"error_string":"influxdb python module not found","module_options":{"batch_size":{"name":"batch_size","type":"int","level":"advanced","flags":0,"default_value":"5000","min":"","max":"","enum_allowed":[],"desc":"How big batches of data points should be when sending to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"database":{"name":"database","type":"str","level":"advanced","flags":0,"default_value":"ceph","min":"","max":"","enum_allowed":[],"desc":"InfluxDB database name. You will need to create this database and grant write privileges to the configured username or the username must have admin privileges to create it.","long_desc":"","tags":[],"see_also":[]},"hostname":{"name":"hostname","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server hostname","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"30","min":"5","max":"","enum_allowed":[],"desc":"Time between reports to InfluxDB. Default 30 seconds.","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"password":{"name":"password","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"password of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"port":{"name":"port","type":"int","level":"advanced","flags":0,"default_value":"8086","min":"","max":"","enum_allowed":[],"desc":"InfluxDB server port","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"ssl":{"name":"ssl","type":"str","level":"advanced","flags":0,"default_value":"false","min":"","max":"","enum_allowed":[],"desc":"Use https connection for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]},"threads":{"name":"threads","type":"int","level":"advanced","flags":0,"default_value":"5","min":"1","max":"32","enum_allowed":[],"desc":"How many worker threads should be spawned for sending data to InfluxDB.","long_desc":"","tags":[],"see_also":[]},"username":{"name":"username","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"username of InfluxDB server user","long_desc":"","tags":[],"see_also":[]},"verify_ssl":{"name":"verify_ssl","type":"str","level":"advanced","flags":0,"default_value":"true","min":"","max":"","enum_allowed":[],"desc":"Verify https cert for InfluxDB server. Use \"true\" or \"false\".","long_desc":"","tags":[],"see_also":[]}}},{"name":"insights","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"iostat","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"k8sevents","can_run":true,"error_string":"","module_options":{"ceph_event_retention_days":{"name":"ceph_event_retention_days","type":"int","level":"advanced","flags":0,"default_value":"7","min":"","max":"","enum_allowed":[],"desc":"Days to hold ceph event information within local cache","long_desc":"","tags":[],"see_also":[]},"config_check_secs":{"name":"config_check_secs","type":"int","level":"advanced","flags":0,"default_value":"10","min":"10","max":"","enum_allowed":[],"desc":"interval (secs) to check for cluster configuration changes","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"localpool","can_run":true,"error_string":"","module_options":{"failure_domain":{"name":"failure_domain","type":"str","level":"advanced","flags":1,"default_value":"host","min":"","max":"","enum_allowed":[],"desc":"failure domain for any created local pool","long_desc":"what failure domain we should separate data replicas across.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"min_size":{"name":"min_size","type":"int","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"default min_size for any created local pool","long_desc":"value to set min_size to (unchanged from Ceph's default if this option is not set)","tags":[],"see_also":[]},"num_rep":{"name":"num_rep","type":"int","level":"advanced","flags":1,"default_value":"3","min":"","max":"","enum_allowed":[],"desc":"default replica count for any created local pool","long_desc":"","tags":[],"see_also":[]},"pg_num":{"name":"pg_num","type":"int","level":"advanced","flags":1,"default_value":"128","min":"","max":"","enum_allowed":[],"desc":"default pg_num for any created local pool","long_desc":"","tags":[],"see_also":[]},"prefix":{"name":"prefix","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"name prefix for any created local pool","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"subtree":{"name":"subtree","type":"str","level":"advanced","flags":1,"default_value":"rack","min":"","max":"","enum_allowed":[],"desc":"CRUSH level for which to create a local pool","long_desc":"which CRUSH subtree type the module should create a pool for.","tags":[],"see_also":[]}}},{"name":"mds_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"mirroring","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"nfs","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"orchestrator","can_run":true,"error_string":"","module_options":{"fail_fs":{"name":"fail_fs","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Fail filesystem for rapid multi-rank mds upgrade","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"orchestrator":{"name":"orchestrator","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["cephadm","rook","test_orchestrator"],"desc":"Orchestrator backend","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_perf_query","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"osd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"pg_autoscaler","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":0,"default_value":"60","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"threshold":{"name":"threshold","type":"float","level":"advanced","flags":0,"default_value":"3.0","min":"1.0","max":"","enum_allowed":[],"desc":"scaling threshold","long_desc":"The factor by which the `NEW PG_NUM` must vary from the current`PG_NUM` before being accepted. Cannot be less than 1.0","tags":[],"see_also":[]}}},{"name":"progress","can_run":true,"error_string":"","module_options":{"allow_pg_recovery_event":{"name":"allow_pg_recovery_event","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow the module to show pg recovery progress","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_completed_events":{"name":"max_completed_events","type":"int","level":"advanced","flags":1,"default_value":"50","min":"","max":"","enum_allowed":[],"desc":"number of past completed events to remember","long_desc":"","tags":[],"see_also":[]},"sleep_interval":{"name":"sleep_interval","type":"secs","level":"advanced","flags":1,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"how long the module is going to sleep","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"prometheus","can_run":true,"error_string":"","module_options":{"cache":{"name":"cache","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"exclude_perf_counters":{"name":"exclude_perf_counters","type":"bool","level":"advanced","flags":1,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Do not include perf-counters in the metrics output","long_desc":"Gathering perf-counters from a single Prometheus exporter can degrade ceph-mgr performance, especially in large clusters. Instead, Ceph-exporter daemons are now used by default for perf-counter gathering. This should only be disabled when no ceph-exporters are deployed.","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools":{"name":"rbd_stats_pools","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rbd_stats_pools_refresh_interval":{"name":"rbd_stats_pools_refresh_interval","type":"int","level":"advanced","flags":0,"default_value":"300","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"scrape_interval":{"name":"scrape_interval","type":"float","level":"advanced","flags":0,"default_value":"15.0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"server_addr":{"name":"server_addr","type":"str","level":"advanced","flags":0,"default_value":"::","min":"","max":"","enum_allowed":[],"desc":"the IPv4 or IPv6 address on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"server_port":{"name":"server_port","type":"int","level":"advanced","flags":1,"default_value":"9283","min":"","max":"","enum_allowed":[],"desc":"the port on which the module listens for HTTP requests","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"stale_cache_strategy":{"name":"stale_cache_strategy","type":"str","level":"advanced","flags":0,"default_value":"log","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_behaviour":{"name":"standby_behaviour","type":"str","level":"advanced","flags":1,"default_value":"default","min":"","max":"","enum_allowed":["default","error"],"desc":"","long_desc":"","tags":[],"see_also":[]},"standby_error_status_code":{"name":"standby_error_status_code","type":"int","level":"advanced","flags":1,"default_value":"500","min":"400","max":"599","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rbd_support","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_snap_create":{"name":"max_concurrent_snap_create","type":"int","level":"advanced","flags":0,"default_value":"10","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"mirror_snapshot_schedule":{"name":"mirror_snapshot_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"trash_purge_schedule":{"name":"trash_purge_schedule","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rgw","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"secondary_zone_period_retry_limit":{"name":"secondary_zone_period_retry_limit","type":"int","level":"advanced","flags":0,"default_value":"5","min":"","max":"","enum_allowed":[],"desc":"RGW module period update retry limit for secondary site","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"rook","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"prometheus_tls_secret_name":{"name":"prometheus_tls_secret_name","type":"str","level":"advanced","flags":0,"default_value":"rook-ceph-prometheus-server-tls","min":"","max":"","enum_allowed":[],"desc":"name of tls secret in k8s for prometheus","long_desc":"","tags":[],"see_also":[]},"secure_monitoring_stack":{"name":"secure_monitoring_stack","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Enable TLS security for all the monitoring stack daemons","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"storage_class":{"name":"storage_class","type":"str","level":"advanced","flags":0,"default_value":"local","min":"","max":"","enum_allowed":[],"desc":"storage class name for LSO-discovered PVs","long_desc":"","tags":[],"see_also":[]}}},{"name":"selftest","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption1":{"name":"roption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"roption2":{"name":"roption2","type":"str","level":"advanced","flags":0,"default_value":"xyz","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption1":{"name":"rwoption1","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption2":{"name":"rwoption2","type":"int","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption3":{"name":"rwoption3","type":"float","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption4":{"name":"rwoption4","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption5":{"name":"rwoption5","type":"bool","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption6":{"name":"rwoption6","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"rwoption7":{"name":"rwoption7","type":"int","level":"advanced","flags":0,"default_value":"","min":"1","max":"42","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testkey":{"name":"testkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testlkey":{"name":"testlkey","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"testnewline":{"name":"testnewline","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"smb","can_run":true,"error_string":"","module_options":{"internal_store_backend":{"name":"internal_store_backend","type":"str","level":"dev","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"set internal store backend. for develoment and testing only","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"update_orchestration":{"name":"update_orchestration","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"automatically update orchestration when smb resources are changed","long_desc":"","tags":[],"see_also":[]}}},{"name":"snap_schedule","can_run":true,"error_string":"","module_options":{"allow_m_granularity":{"name":"allow_m_granularity","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"allow minute scheduled snapshots","long_desc":"","tags":[],"see_also":[]},"dump_on_update":{"name":"dump_on_update","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"dump database to debug log on update","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"stats","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"status","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telegraf","can_run":true,"error_string":"","module_options":{"address":{"name":"address","type":"str","level":"advanced","flags":0,"default_value":"unixgram:///tmp/telegraf.sock","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"secs","level":"advanced","flags":0,"default_value":"15","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"telemetry","can_run":true,"error_string":"","module_options":{"channel_basic":{"name":"channel_basic","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share basic cluster information (size, version)","long_desc":"","tags":[],"see_also":[]},"channel_crash":{"name":"channel_crash","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share metadata about Ceph daemon crashes (version, stack straces, etc)","long_desc":"","tags":[],"see_also":[]},"channel_device":{"name":"channel_device","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Share device health metrics (e.g., SMART data, minus potentially identifying info like serial numbers)","long_desc":"","tags":[],"see_also":[]},"channel_ident":{"name":"channel_ident","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share a user-provided description and/or contact email for the cluster","long_desc":"","tags":[],"see_also":[]},"channel_perf":{"name":"channel_perf","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Share various performance metrics of a cluster","long_desc":"","tags":[],"see_also":[]},"contact":{"name":"contact","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"description":{"name":"description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"device_url":{"name":"device_url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/device","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"enabled":{"name":"enabled","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"interval":{"name":"interval","type":"int","level":"advanced","flags":0,"default_value":"24","min":"8","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"last_opt_revision":{"name":"last_opt_revision","type":"int","level":"advanced","flags":0,"default_value":"1","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard":{"name":"leaderboard","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"leaderboard_description":{"name":"leaderboard_description","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"organization":{"name":"organization","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"proxy":{"name":"proxy","type":"str","level":"advanced","flags":0,"default_value":"","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"url":{"name":"url","type":"str","level":"advanced","flags":0,"default_value":"https://telemetry.ceph.com/report","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"test_orchestrator","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}},{"name":"volumes","can_run":true,"error_string":"","module_options":{"log_level":{"name":"log_level","type":"str","level":"advanced","flags":1,"default_value":"","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster":{"name":"log_to_cluster","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_cluster_level":{"name":"log_to_cluster_level","type":"str","level":"advanced","flags":1,"default_value":"info","min":"","max":"","enum_allowed":["","critical","debug","error","info","warning"],"desc":"","long_desc":"","tags":[],"see_also":[]},"log_to_file":{"name":"log_to_file","type":"bool","level":"advanced","flags":1,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]},"max_concurrent_clones":{"name":"max_concurrent_clones","type":"int","level":"advanced","flags":0,"default_value":"4","min":"","max":"","enum_allowed":[],"desc":"Number of asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_cloning":{"name":"pause_cloning","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous cloner threads","long_desc":"","tags":[],"see_also":[]},"pause_purging":{"name":"pause_purging","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Pause asynchronous subvolume purge threads","long_desc":"","tags":[],"see_also":[]},"periodic_async_work":{"name":"periodic_async_work","type":"bool","level":"advanced","flags":0,"default_value":"False","min":"","max":"","enum_allowed":[],"desc":"Periodically check for async work","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_delay":{"name":"snapshot_clone_delay","type":"int","level":"advanced","flags":0,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"Delay clone begin operation by snapshot_clone_delay seconds","long_desc":"","tags":[],"see_also":[]},"snapshot_clone_no_wait":{"name":"snapshot_clone_no_wait","type":"bool","level":"advanced","flags":0,"default_value":"True","min":"","max":"","enum_allowed":[],"desc":"Reject subvolume clone request when cloner threads are busy","long_desc":"","tags":[],"see_also":[]},"sqlite3_killpoint":{"name":"sqlite3_killpoint","type":"int","level":"dev","flags":1,"default_value":"0","min":"","max":"","enum_allowed":[],"desc":"","long_desc":"","tags":[],"see_also":[]}}}],"services":{"dashboard":"https://192.168.123.102:8443/"},"always_on_modules":{"octopus":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"pacific":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"quincy":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"reef":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"squid":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"],"tentacle":["balancer","crash","devicehealth","orchestrator","pg_autoscaler","progress","rbd_support","status","telemetry","volumes"]},"force_disabled_modules":{},"last_failure_osd_epoch":3,"active_clients":[{"name":"devicehealth","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":2960834819}]},{"name":"libcephsqlite","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":1664936117}]},{"name":"rbd_support","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":2742290181}]},{"name":"volumes","addrvec":[{"type":"v2","addr":"192.168.123.102:0","nonce":241302811}]}]} 2026-03-21T06:55:57.328 INFO:tasks.cephadm.ceph_manager.ceph:mgr available! 2026-03-21T06:55:57.328 INFO:tasks.cephadm.ceph_manager.ceph:waiting for all up 2026-03-21T06:55:57.328 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd dump --format=json 2026-03-21T06:55:57.330 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:57.330 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:55:57 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:57.330 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:55:57 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:57.331 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:55:57 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:55:57.593 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:57.609 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: cephadm 2026-03-21T06:55:56.318744+0000 mgr.x (mgr.14152) 343 : cephadm [INF] Deploying daemon iscsi.iscsi.b on vm07 2026-03-21T06:55:57.609 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: cephadm 2026-03-21T06:55:56.318744+0000 mgr.x (mgr.14152) 343 : cephadm [INF] Deploying daemon iscsi.iscsi.b on vm07 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:56.496926+0000 mon.a (mon.0) 773 : audit [DBG] from='client.? 192.168.123.102:0/2862339535' entity='client.admin' cmd={"prefix": "config log", "num": 1, "format": "json"} : dispatch 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:56.496926+0000 mon.a (mon.0) 773 : audit [DBG] from='client.? 192.168.123.102:0/2862339535' entity='client.admin' cmd={"prefix": "config log", "num": 1, "format": "json"} : dispatch 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:56.810661+0000 mon.c (mon.1) 29 : audit [DBG] from='client.? 192.168.123.102:0/669926853' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:56.810661+0000 mon.c (mon.1) 29 : audit [DBG] from='client.? 192.168.123.102:0/669926853' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.052112+0000 mon.a (mon.0) 774 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"} : dispatch 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.052112+0000 mon.a (mon.0) 774 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"} : dispatch 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.169298+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.169298+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.180665+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.180665+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.186196+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.186196+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.199380+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.199380+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.213290+0000 mon.a (mon.0) 779 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.213290+0000 mon.a (mon.0) 779 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.242219+0000 mon.a (mon.0) 780 : audit [DBG] from='client.? 192.168.123.102:0/754580631' entity='client.admin' cmd={"prefix": "mgr dump", "format": "json"} : dispatch 2026-03-21T06:55:57.610 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[17657]: audit 2026-03-21T06:55:57.242219+0000 mon.a (mon.0) 780 : audit [DBG] from='client.? 192.168.123.102:0/754580631' entity='client.admin' cmd={"prefix": "mgr dump", "format": "json"} : dispatch 2026-03-21T06:55:57.610 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[49158]: debug Successfully removed blocklist entry 2026-03-21T06:55:57.610 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:55:57 vm02 bash[49158]: debug Removing blocklisted entry for this host : 192.168.123.102:6800/4266123507 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: cephadm 2026-03-21T06:55:56.318744+0000 mgr.x (mgr.14152) 343 : cephadm [INF] Deploying daemon iscsi.iscsi.b on vm07 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: cephadm 2026-03-21T06:55:56.318744+0000 mgr.x (mgr.14152) 343 : cephadm [INF] Deploying daemon iscsi.iscsi.b on vm07 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:56.496926+0000 mon.a (mon.0) 773 : audit [DBG] from='client.? 192.168.123.102:0/2862339535' entity='client.admin' cmd={"prefix": "config log", "num": 1, "format": "json"} : dispatch 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:56.496926+0000 mon.a (mon.0) 773 : audit [DBG] from='client.? 192.168.123.102:0/2862339535' entity='client.admin' cmd={"prefix": "config log", "num": 1, "format": "json"} : dispatch 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:56.810661+0000 mon.c (mon.1) 29 : audit [DBG] from='client.? 192.168.123.102:0/669926853' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:56.810661+0000 mon.c (mon.1) 29 : audit [DBG] from='client.? 192.168.123.102:0/669926853' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.052112+0000 mon.a (mon.0) 774 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"} : dispatch 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.052112+0000 mon.a (mon.0) 774 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"} : dispatch 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.169298+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.169298+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.180665+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.180665+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.186196+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.186196+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.199380+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.199380+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.213290+0000 mon.a (mon.0) 779 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.213290+0000 mon.a (mon.0) 779 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.242219+0000 mon.a (mon.0) 780 : audit [DBG] from='client.? 192.168.123.102:0/754580631' entity='client.admin' cmd={"prefix": "mgr dump", "format": "json"} : dispatch 2026-03-21T06:55:57.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:57 vm07 bash[19945]: audit 2026-03-21T06:55:57.242219+0000 mon.a (mon.0) 780 : audit [DBG] from='client.? 192.168.123.102:0/754580631' entity='client.admin' cmd={"prefix": "mgr dump", "format": "json"} : dispatch 2026-03-21T06:55:57.651 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: debug Processing osd blocklist entries for this node 2026-03-21T06:55:57.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: cephadm 2026-03-21T06:55:56.318744+0000 mgr.x (mgr.14152) 343 : cephadm [INF] Deploying daemon iscsi.iscsi.b on vm07 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: cephadm 2026-03-21T06:55:56.318744+0000 mgr.x (mgr.14152) 343 : cephadm [INF] Deploying daemon iscsi.iscsi.b on vm07 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:56.496926+0000 mon.a (mon.0) 773 : audit [DBG] from='client.? 192.168.123.102:0/2862339535' entity='client.admin' cmd={"prefix": "config log", "num": 1, "format": "json"} : dispatch 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:56.496926+0000 mon.a (mon.0) 773 : audit [DBG] from='client.? 192.168.123.102:0/2862339535' entity='client.admin' cmd={"prefix": "config log", "num": 1, "format": "json"} : dispatch 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:56.810661+0000 mon.c (mon.1) 29 : audit [DBG] from='client.? 192.168.123.102:0/669926853' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:56.810661+0000 mon.c (mon.1) 29 : audit [DBG] from='client.? 192.168.123.102:0/669926853' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.052112+0000 mon.a (mon.0) 774 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"} : dispatch 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.052112+0000 mon.a (mon.0) 774 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"} : dispatch 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.169298+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.169298+0000 mon.a (mon.0) 775 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.180665+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.180665+0000 mon.a (mon.0) 776 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.186196+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.186196+0000 mon.a (mon.0) 777 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.199380+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.199380+0000 mon.a (mon.0) 778 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.213290+0000 mon.a (mon.0) 779 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.213290+0000 mon.a (mon.0) 779 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.242219+0000 mon.a (mon.0) 780 : audit [DBG] from='client.? 192.168.123.102:0/754580631' entity='client.admin' cmd={"prefix": "mgr dump", "format": "json"} : dispatch 2026-03-21T06:55:57.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:57 vm04 bash[20194]: audit 2026-03-21T06:55:57.242219+0000 mon.a (mon.0) 780 : audit [DBG] from='client.? 192.168.123.102:0/754580631' entity='client.admin' cmd={"prefix": "mgr dump", "format": "json"} : dispatch 2026-03-21T06:55:57.923 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:55:57.923 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":60,"fsid":"b16ecafc-24f1-11f1-8ede-8330751617ee","created":"2026-03-21T06:47:47.924264+0000","modified":"2026-03-21T06:55:57.468840+0000","last_up_change":"2026-03-21T06:55:49.834055+0000","last_in_change":"2026-03-21T06:53:38.245564+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":23,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-21T06:51:07.780816+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"24","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"datapool","create_time":"2026-03-21T06:55:49.423058+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":3,"pg_placement_num":3,"pg_placement_num_target":3,"pg_num_target":3,"pg_num_pending":3,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"59","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":59,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":5.3299999237060547,"score_stable":5.3299999237060547,"optimal_score":0.75,"raw_score_acting":4,"raw_score_stable":4,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"5a293892-6a20-454b-9116-106cddf14695","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":8,"last_clean_end":9,"up_from":12,"up_thru":27,"down_at":11,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6803","nonce":2802530539}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6805","nonce":2802530539}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6809","nonce":2802530539}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6807","nonce":2802530539}]},"public_addr":"192.168.123.102:6803/2802530539","cluster_addr":"192.168.123.102:6805/2802530539","heartbeat_back_addr":"192.168.123.102:6809/2802530539","heartbeat_front_addr":"192.168.123.102:6807/2802530539","state":["exists","up"]},{"osd":1,"uuid":"38235c66-ddf2-470c-8f4a-dac1efac5f75","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":16,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6811","nonce":2868931959}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6813","nonce":2868931959}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6817","nonce":2868931959}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6815","nonce":2868931959}]},"public_addr":"192.168.123.102:6811/2868931959","cluster_addr":"192.168.123.102:6813/2868931959","heartbeat_back_addr":"192.168.123.102:6817/2868931959","heartbeat_front_addr":"192.168.123.102:6815/2868931959","state":["exists","up"]},{"osd":2,"uuid":"2bbd253a-cfb7-4d35-a80c-5ab1d023b91e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":56,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6801","nonce":3750025046}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6803","nonce":3750025046}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6807","nonce":3750025046}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6805","nonce":3750025046}]},"public_addr":"192.168.123.104:6801/3750025046","cluster_addr":"192.168.123.104:6803/3750025046","heartbeat_back_addr":"192.168.123.104:6807/3750025046","heartbeat_front_addr":"192.168.123.104:6805/3750025046","state":["exists","up"]},{"osd":3,"uuid":"c2851fc5-1154-4c44-91cd-287a260b97a7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":56,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6808","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6809","nonce":3282486205}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6810","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6811","nonce":3282486205}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6814","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6815","nonce":3282486205}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6812","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6813","nonce":3282486205}]},"public_addr":"192.168.123.104:6809/3282486205","cluster_addr":"192.168.123.104:6811/3282486205","heartbeat_back_addr":"192.168.123.104:6815/3282486205","heartbeat_front_addr":"192.168.123.104:6813/3282486205","state":["exists","up"]},{"osd":4,"uuid":"22bec0cb-fc4e-4173-bab0-1ea7c238683c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6816","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6817","nonce":3736423955}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6818","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6819","nonce":3736423955}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6822","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6823","nonce":3736423955}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6820","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6821","nonce":3736423955}]},"public_addr":"192.168.123.104:6817/3736423955","cluster_addr":"192.168.123.104:6819/3736423955","heartbeat_back_addr":"192.168.123.104:6823/3736423955","heartbeat_front_addr":"192.168.123.104:6821/3736423955","state":["exists","up"]},{"osd":5,"uuid":"a697bb58-dae5-4cba-b085-8e5769734d61","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":36,"last_clean_end":38,"up_from":41,"up_thru":0,"down_at":39,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6800","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6801","nonce":102330835}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6802","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6803","nonce":102330835}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6806","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6807","nonce":102330835}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6804","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6805","nonce":102330835}]},"public_addr":"192.168.123.107:6801/102330835","cluster_addr":"192.168.123.107:6803/102330835","heartbeat_back_addr":"192.168.123.107:6807/102330835","heartbeat_front_addr":"192.168.123.107:6805/102330835","state":["exists","up"]},{"osd":6,"uuid":"7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":45,"last_clean_end":46,"up_from":49,"up_thru":0,"down_at":47,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6808","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6809","nonce":2090727328}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6810","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6811","nonce":2090727328}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6814","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6815","nonce":2090727328}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6812","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6813","nonce":2090727328}]},"public_addr":"192.168.123.107:6809/2090727328","cluster_addr":"192.168.123.107:6811/2090727328","heartbeat_back_addr":"192.168.123.107:6815/2090727328","heartbeat_front_addr":"192.168.123.107:6813/2090727328","state":["exists","up"]},{"osd":7,"uuid":"8e806dd2-e80e-40ab-8996-ff252b7491d9","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":52,"last_clean_end":53,"up_from":56,"up_thru":0,"down_at":54,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6816","nonce":199307136},{"type":"v1","addr":"192.168.123.107:6817","nonce":199307136}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6818","nonce":199307136},{"type":"v1","addr":"192.168.123.107:6819","nonce":199307136}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6822","nonce":199307136},{"type":"v1","addr":"192.168.123.107:6823","nonce":199307136}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6820","nonce":199307136},{"type":"v1","addr":"192.168.123.107:6821","nonce":199307136}]},"public_addr":"192.168.123.107:6817/199307136","cluster_addr":"192.168.123.107:6819/199307136","heartbeat_back_addr":"192.168.123.107:6823/199307136","heartbeat_front_addr":"192.168.123.107:6821/199307136","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"2026-03-21T06:50:52.082988+0000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:48:49.774464+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:50:58.549660+0000","dead_epoch":15},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:05.687654+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:13.735249+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:20.945258+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:28.648539+0000","dead_epoch":39},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:53:37.274136+0000","dead_epoch":46},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:53:47.648093+0000","dead_epoch":54}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:6800/4266123507":"2026-03-22T06:48:09.528930+0000","192.168.123.102:0/1909593542":"2026-03-22T06:48:09.528930+0000","192.168.123.102:0/1252995422":"2026-03-22T06:48:09.528930+0000","192.168.123.102:0/1580070482":"2026-03-22T06:48:09.528930+0000","192.168.123.102:6801/260656842":"2026-03-22T06:47:58.527875+0000","192.168.123.102:6800/260656842":"2026-03-22T06:47:58.527875+0000","192.168.123.102:0/2023406208":"2026-03-22T06:47:58.527875+0000","192.168.123.102:0/1804345957":"2026-03-22T06:47:58.527875+0000","192.168.123.102:0/288488007":"2026-03-22T06:47:58.527875+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: debug Reading the configuration object to update local LIO configuration 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: debug Configuration does not have an entry for this host(vm07.local) - nothing to define to LIO 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: * Serving Flask app 'rbd-target-api' (lazy loading) 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: * Environment: production 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: Use a production WSGI server instead. 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: * Debug mode: off 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: debug * Running on all addresses. 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: * Running on all addresses. 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: debug * Running on http://[::1]:5000/ (Press CTRL+C to quit) 2026-03-21T06:55:57.932 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:55:57 vm07 bash[42142]: * Running on http://[::1]:5000/ (Press CTRL+C to quit) 2026-03-21T06:55:57.973 INFO:tasks.cephadm.ceph_manager.ceph:all up! 2026-03-21T06:55:57.973 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd dump --format=json 2026-03-21T06:55:58.207 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:58.524 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:55:58.524 INFO:teuthology.orchestra.run.vm02.stdout:{"epoch":61,"fsid":"b16ecafc-24f1-11f1-8ede-8330751617ee","created":"2026-03-21T06:47:47.924264+0000","modified":"2026-03-21T06:55:58.489598+0000","last_up_change":"2026-03-21T06:55:49.834055+0000","last_in_change":"2026-03-21T06:53:38.245564+0000","flags":"sortbitwise,recovery_deletes,purged_snapdirs,pglog_hardlimit","flags_num":5799936,"flags_set":["pglog_hardlimit","purged_snapdirs","recovery_deletes","sortbitwise"],"crush_version":23,"full_ratio":0.94999998807907104,"backfillfull_ratio":0.89999997615814209,"nearfull_ratio":0.85000002384185791,"cluster_snapshot":"","pool_max":2,"max_osd":8,"require_min_compat_client":"luminous","min_compat_client":"jewel","require_osd_release":"tentacle","allow_crimson":false,"pools":[{"pool":1,"pool_name":".mgr","create_time":"2026-03-21T06:51:07.780816+0000","flags":1,"flags_names":"hashpspool","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":1,"pg_placement_num":1,"pg_placement_num_target":1,"pg_num_target":1,"pg_num_pending":1,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"24","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":0,"snap_epoch":0,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{"pg_num_max":32,"pg_num_min":1},"application_metadata":{"mgr":{}},"read_balance":{"score_type":"Fair distribution","score_acting":7.8899998664855957,"score_stable":7.8899998664855957,"optimal_score":0.37999999523162842,"raw_score_acting":3,"raw_score_stable":3,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}},{"pool":2,"pool_name":"datapool","create_time":"2026-03-21T06:55:49.423058+0000","flags":8193,"flags_names":"hashpspool,selfmanaged_snaps","type":1,"size":3,"min_size":2,"crush_rule":0,"peering_crush_bucket_count":0,"peering_crush_bucket_target":0,"peering_crush_bucket_barrier":0,"peering_crush_bucket_mandatory_member":2147483647,"is_stretch_pool":false,"object_hash":2,"pg_autoscale_mode":"off","pg_num":3,"pg_placement_num":3,"pg_placement_num_target":3,"pg_num_target":3,"pg_num_pending":3,"last_pg_merge_meta":{"source_pgid":"0.0","ready_epoch":0,"last_epoch_started":0,"last_epoch_clean":0,"source_version":"0'0","target_version":"0'0"},"last_change":"59","last_force_op_resend":"0","last_force_op_resend_prenautilus":"0","last_force_op_resend_preluminous":"0","auid":0,"snap_mode":"selfmanaged","snap_seq":2,"snap_epoch":59,"pool_snaps":[],"removed_snaps":"[]","quota_max_bytes":0,"quota_max_objects":0,"tiers":[],"tier_of":-1,"read_tier":-1,"write_tier":-1,"cache_mode":"none","target_max_bytes":0,"target_max_objects":0,"cache_target_dirty_ratio_micro":400000,"cache_target_dirty_high_ratio_micro":600000,"cache_target_full_ratio_micro":800000,"cache_min_flush_age":0,"cache_min_evict_age":0,"erasure_code_profile":"","hit_set_params":{"type":"none"},"hit_set_period":0,"hit_set_count":0,"use_gmt_hitset":true,"min_read_recency_for_promote":0,"min_write_recency_for_promote":0,"hit_set_grade_decay_rate":0,"hit_set_search_last_n":0,"grade_table":[],"stripe_width":0,"expected_num_objects":0,"fast_read":false,"nonprimary_shards":"{}","options":{},"application_metadata":{"rbd":{}},"read_balance":{"score_type":"Fair distribution","score_acting":5.3299999237060547,"score_stable":5.3299999237060547,"optimal_score":0.75,"raw_score_acting":4,"raw_score_stable":4,"primary_affinity_weighted":1,"average_primary_affinity":1,"average_primary_affinity_weighted":1}}],"osds":[{"osd":0,"uuid":"5a293892-6a20-454b-9116-106cddf14695","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":8,"last_clean_end":9,"up_from":12,"up_thru":27,"down_at":11,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6802","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6803","nonce":2802530539}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6804","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6805","nonce":2802530539}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6808","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6809","nonce":2802530539}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6806","nonce":2802530539},{"type":"v1","addr":"192.168.123.102:6807","nonce":2802530539}]},"public_addr":"192.168.123.102:6803/2802530539","cluster_addr":"192.168.123.102:6805/2802530539","heartbeat_back_addr":"192.168.123.102:6809/2802530539","heartbeat_front_addr":"192.168.123.102:6807/2802530539","state":["exists","up"]},{"osd":1,"uuid":"38235c66-ddf2-470c-8f4a-dac1efac5f75","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":18,"up_thru":0,"down_at":16,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6810","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6811","nonce":2868931959}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6812","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6813","nonce":2868931959}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6816","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6817","nonce":2868931959}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.102:6814","nonce":2868931959},{"type":"v1","addr":"192.168.123.102:6815","nonce":2868931959}]},"public_addr":"192.168.123.102:6811/2868931959","cluster_addr":"192.168.123.102:6813/2868931959","heartbeat_back_addr":"192.168.123.102:6817/2868931959","heartbeat_front_addr":"192.168.123.102:6815/2868931959","state":["exists","up"]},{"osd":2,"uuid":"2bbd253a-cfb7-4d35-a80c-5ab1d023b91e","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":21,"up_thru":56,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6800","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6801","nonce":3750025046}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6802","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6803","nonce":3750025046}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6806","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6807","nonce":3750025046}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6804","nonce":3750025046},{"type":"v1","addr":"192.168.123.104:6805","nonce":3750025046}]},"public_addr":"192.168.123.104:6801/3750025046","cluster_addr":"192.168.123.104:6803/3750025046","heartbeat_back_addr":"192.168.123.104:6807/3750025046","heartbeat_front_addr":"192.168.123.104:6805/3750025046","state":["exists","up"]},{"osd":3,"uuid":"c2851fc5-1154-4c44-91cd-287a260b97a7","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":27,"up_thru":56,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6808","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6809","nonce":3282486205}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6810","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6811","nonce":3282486205}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6814","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6815","nonce":3282486205}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6812","nonce":3282486205},{"type":"v1","addr":"192.168.123.104:6813","nonce":3282486205}]},"public_addr":"192.168.123.104:6809/3282486205","cluster_addr":"192.168.123.104:6811/3282486205","heartbeat_back_addr":"192.168.123.104:6815/3282486205","heartbeat_front_addr":"192.168.123.104:6813/3282486205","state":["exists","up"]},{"osd":4,"uuid":"22bec0cb-fc4e-4173-bab0-1ea7c238683c","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":0,"last_clean_end":0,"up_from":32,"up_thru":0,"down_at":0,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6816","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6817","nonce":3736423955}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6818","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6819","nonce":3736423955}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6822","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6823","nonce":3736423955}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.104:6820","nonce":3736423955},{"type":"v1","addr":"192.168.123.104:6821","nonce":3736423955}]},"public_addr":"192.168.123.104:6817/3736423955","cluster_addr":"192.168.123.104:6819/3736423955","heartbeat_back_addr":"192.168.123.104:6823/3736423955","heartbeat_front_addr":"192.168.123.104:6821/3736423955","state":["exists","up"]},{"osd":5,"uuid":"a697bb58-dae5-4cba-b085-8e5769734d61","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":36,"last_clean_end":38,"up_from":41,"up_thru":0,"down_at":39,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6800","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6801","nonce":102330835}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6802","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6803","nonce":102330835}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6806","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6807","nonce":102330835}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6804","nonce":102330835},{"type":"v1","addr":"192.168.123.107:6805","nonce":102330835}]},"public_addr":"192.168.123.107:6801/102330835","cluster_addr":"192.168.123.107:6803/102330835","heartbeat_back_addr":"192.168.123.107:6807/102330835","heartbeat_front_addr":"192.168.123.107:6805/102330835","state":["exists","up"]},{"osd":6,"uuid":"7ca50915-6ca3-4c9a-8e2b-1ba9d600c9be","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":45,"last_clean_end":46,"up_from":49,"up_thru":0,"down_at":47,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6808","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6809","nonce":2090727328}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6810","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6811","nonce":2090727328}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6814","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6815","nonce":2090727328}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6812","nonce":2090727328},{"type":"v1","addr":"192.168.123.107:6813","nonce":2090727328}]},"public_addr":"192.168.123.107:6809/2090727328","cluster_addr":"192.168.123.107:6811/2090727328","heartbeat_back_addr":"192.168.123.107:6815/2090727328","heartbeat_front_addr":"192.168.123.107:6813/2090727328","state":["exists","up"]},{"osd":7,"uuid":"8e806dd2-e80e-40ab-8996-ff252b7491d9","up":1,"in":1,"weight":1,"primary_affinity":1,"last_clean_begin":52,"last_clean_end":53,"up_from":56,"up_thru":0,"down_at":54,"lost_at":0,"public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6816","nonce":199307136},{"type":"v1","addr":"192.168.123.107:6817","nonce":199307136}]},"cluster_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6818","nonce":199307136},{"type":"v1","addr":"192.168.123.107:6819","nonce":199307136}]},"heartbeat_back_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6822","nonce":199307136},{"type":"v1","addr":"192.168.123.107:6823","nonce":199307136}]},"heartbeat_front_addrs":{"addrvec":[{"type":"v2","addr":"192.168.123.107:6820","nonce":199307136},{"type":"v1","addr":"192.168.123.107:6821","nonce":199307136}]},"public_addr":"192.168.123.107:6817/199307136","cluster_addr":"192.168.123.107:6819/199307136","heartbeat_back_addr":"192.168.123.107:6823/199307136","heartbeat_front_addr":"192.168.123.107:6821/199307136","state":["exists","up"]}],"osd_xinfo":[{"osd":0,"down_stamp":"2026-03-21T06:50:52.082988+0000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:48:49.774464+0000","dead_epoch":0},{"osd":1,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:50:58.549660+0000","dead_epoch":15},{"osd":2,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:05.687654+0000","dead_epoch":0},{"osd":3,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:13.735249+0000","dead_epoch":0},{"osd":4,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:20.945258+0000","dead_epoch":0},{"osd":5,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:51:28.648539+0000","dead_epoch":39},{"osd":6,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:53:37.274136+0000","dead_epoch":46},{"osd":7,"down_stamp":"0.000000","laggy_probability":0,"laggy_interval":0,"features":4544132024016699391,"old_weight":0,"last_purged_snaps_scrub":"2026-03-21T06:53:47.648093+0000","dead_epoch":54}],"pg_upmap":[],"pg_upmap_items":[],"pg_upmap_primaries":[],"pg_temp":[],"primary_temp":[],"blocklist":{"192.168.123.102:0/1909593542":"2026-03-22T06:48:09.528930+0000","192.168.123.102:0/1252995422":"2026-03-22T06:48:09.528930+0000","192.168.123.102:0/1580070482":"2026-03-22T06:48:09.528930+0000","192.168.123.102:6801/260656842":"2026-03-22T06:47:58.527875+0000","192.168.123.102:6800/260656842":"2026-03-22T06:47:58.527875+0000","192.168.123.102:0/2023406208":"2026-03-22T06:47:58.527875+0000","192.168.123.102:0/1804345957":"2026-03-22T06:47:58.527875+0000","192.168.123.102:0/288488007":"2026-03-22T06:47:58.527875+0000"},"range_blocklist":{},"erasure_code_profiles":{"default":{"crush-failure-domain":"osd","k":"2","m":"1","plugin":"isa","technique":"reed_sol_van"}},"removed_snaps_queue":[{"pool":2,"snaps":[{"begin":2,"length":1}]}],"new_removed_snaps":[],"new_purged_snaps":[],"crush_node_flags":{},"device_class_flags":{},"stretch_mode":{"stretch_mode_enabled":false,"stretch_bucket_count":0,"degraded_stretch_mode":0,"recovering_stretch_mode":0,"stretch_mode_bucket":0}} 2026-03-21T06:55:58.582 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph tell osd.0 flush_pg_stats 2026-03-21T06:55:58.582 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph tell osd.1 flush_pg_stats 2026-03-21T06:55:58.582 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph tell osd.2 flush_pg_stats 2026-03-21T06:55:58.582 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph tell osd.3 flush_pg_stats 2026-03-21T06:55:58.582 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph tell osd.4 flush_pg_stats 2026-03-21T06:55:58.582 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph tell osd.5 flush_pg_stats 2026-03-21T06:55:58.582 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph tell osd.6 flush_pg_stats 2026-03-21T06:55:58.583 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph tell osd.7 flush_pg_stats 2026-03-21T06:55:58.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: cephadm 2026-03-21T06:55:57.186665+0000 mgr.x (mgr.14152) 344 : cephadm [INF] Checking pool "datapool" exists for service iscsi.datapool 2026-03-21T06:55:58.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: cephadm 2026-03-21T06:55:57.186665+0000 mgr.x (mgr.14152) 344 : cephadm [INF] Checking pool "datapool" exists for service iscsi.datapool 2026-03-21T06:55:58.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: audit 2026-03-21T06:55:57.471847+0000 mon.a (mon.0) 781 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"}]': finished 2026-03-21T06:55:58.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: audit 2026-03-21T06:55:57.471847+0000 mon.a (mon.0) 781 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"}]': finished 2026-03-21T06:55:58.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: cluster 2026-03-21T06:55:57.477414+0000 mon.a (mon.0) 782 : cluster [DBG] osdmap e60: 8 total, 8 up, 8 in 2026-03-21T06:55:58.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: cluster 2026-03-21T06:55:57.477414+0000 mon.a (mon.0) 782 : cluster [DBG] osdmap e60: 8 total, 8 up, 8 in 2026-03-21T06:55:58.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: cluster 2026-03-21T06:55:57.604320+0000 mgr.x (mgr.14152) 345 : cluster [DBG] pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:55:58.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: cluster 2026-03-21T06:55:57.604320+0000 mgr.x (mgr.14152) 345 : cluster [DBG] pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:55:58.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: audit 2026-03-21T06:55:57.662084+0000 mon.b (mon.2) 18 : audit [DBG] from='client.? 192.168.123.107:0/3580489592' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:58.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: audit 2026-03-21T06:55:57.662084+0000 mon.b (mon.2) 18 : audit [DBG] from='client.? 192.168.123.107:0/3580489592' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:58.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: audit 2026-03-21T06:55:57.718084+0000 mon.a (mon.0) 783 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"} : dispatch 2026-03-21T06:55:58.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: audit 2026-03-21T06:55:57.718084+0000 mon.a (mon.0) 783 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"} : dispatch 2026-03-21T06:55:58.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: audit 2026-03-21T06:55:57.924535+0000 mon.a (mon.0) 784 : audit [DBG] from='client.? 192.168.123.102:0/773871367' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:58.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:58 vm04 bash[20194]: audit 2026-03-21T06:55:57.924535+0000 mon.a (mon.0) 784 : audit [DBG] from='client.? 192.168.123.102:0/773871367' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: cephadm 2026-03-21T06:55:57.186665+0000 mgr.x (mgr.14152) 344 : cephadm [INF] Checking pool "datapool" exists for service iscsi.datapool 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: cephadm 2026-03-21T06:55:57.186665+0000 mgr.x (mgr.14152) 344 : cephadm [INF] Checking pool "datapool" exists for service iscsi.datapool 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: audit 2026-03-21T06:55:57.471847+0000 mon.a (mon.0) 781 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"}]': finished 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: audit 2026-03-21T06:55:57.471847+0000 mon.a (mon.0) 781 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"}]': finished 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: cluster 2026-03-21T06:55:57.477414+0000 mon.a (mon.0) 782 : cluster [DBG] osdmap e60: 8 total, 8 up, 8 in 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: cluster 2026-03-21T06:55:57.477414+0000 mon.a (mon.0) 782 : cluster [DBG] osdmap e60: 8 total, 8 up, 8 in 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: cluster 2026-03-21T06:55:57.604320+0000 mgr.x (mgr.14152) 345 : cluster [DBG] pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: cluster 2026-03-21T06:55:57.604320+0000 mgr.x (mgr.14152) 345 : cluster [DBG] pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: audit 2026-03-21T06:55:57.662084+0000 mon.b (mon.2) 18 : audit [DBG] from='client.? 192.168.123.107:0/3580489592' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: audit 2026-03-21T06:55:57.662084+0000 mon.b (mon.2) 18 : audit [DBG] from='client.? 192.168.123.107:0/3580489592' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: audit 2026-03-21T06:55:57.718084+0000 mon.a (mon.0) 783 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"} : dispatch 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: audit 2026-03-21T06:55:57.718084+0000 mon.a (mon.0) 783 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"} : dispatch 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: audit 2026-03-21T06:55:57.924535+0000 mon.a (mon.0) 784 : audit [DBG] from='client.? 192.168.123.102:0/773871367' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:58.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[17657]: audit 2026-03-21T06:55:57.924535+0000 mon.a (mon.0) 784 : audit [DBG] from='client.? 192.168.123.102:0/773871367' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:58.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[49158]: debug Successfully removed blocklist entry 2026-03-21T06:55:58.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[49158]: debug Removing blocklisted entry for this host : 192.168.123.102:0/1909593542 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: cephadm 2026-03-21T06:55:57.186665+0000 mgr.x (mgr.14152) 344 : cephadm [INF] Checking pool "datapool" exists for service iscsi.datapool 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: cephadm 2026-03-21T06:55:57.186665+0000 mgr.x (mgr.14152) 344 : cephadm [INF] Checking pool "datapool" exists for service iscsi.datapool 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: audit 2026-03-21T06:55:57.471847+0000 mon.a (mon.0) 781 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"}]': finished 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: audit 2026-03-21T06:55:57.471847+0000 mon.a (mon.0) 781 : audit [INF] from='client.? 192.168.123.102:0/761285100' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/4266123507"}]': finished 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: cluster 2026-03-21T06:55:57.477414+0000 mon.a (mon.0) 782 : cluster [DBG] osdmap e60: 8 total, 8 up, 8 in 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: cluster 2026-03-21T06:55:57.477414+0000 mon.a (mon.0) 782 : cluster [DBG] osdmap e60: 8 total, 8 up, 8 in 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: cluster 2026-03-21T06:55:57.604320+0000 mgr.x (mgr.14152) 345 : cluster [DBG] pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: cluster 2026-03-21T06:55:57.604320+0000 mgr.x (mgr.14152) 345 : cluster [DBG] pgmap v284: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: audit 2026-03-21T06:55:57.662084+0000 mon.b (mon.2) 18 : audit [DBG] from='client.? 192.168.123.107:0/3580489592' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: audit 2026-03-21T06:55:57.662084+0000 mon.b (mon.2) 18 : audit [DBG] from='client.? 192.168.123.107:0/3580489592' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist ls"} : dispatch 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: audit 2026-03-21T06:55:57.718084+0000 mon.a (mon.0) 783 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"} : dispatch 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: audit 2026-03-21T06:55:57.718084+0000 mon.a (mon.0) 783 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"} : dispatch 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: audit 2026-03-21T06:55:57.924535+0000 mon.a (mon.0) 784 : audit [DBG] from='client.? 192.168.123.102:0/773871367' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:58.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:58 vm07 bash[19945]: audit 2026-03-21T06:55:57.924535+0000 mon.a (mon.0) 784 : audit [DBG] from='client.? 192.168.123.102:0/773871367' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:58.943 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:59.107 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:59.141 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:59.144 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:59.185 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:59.192 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:59.237 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[49158]: debug Successfully removed blocklist entry 2026-03-21T06:55:59.237 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:55:58 vm02 bash[49158]: debug Removing blocklisted entry for this host : 192.168.123.102:0/1252995422 2026-03-21T06:55:59.261 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:59.277 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:59.565 INFO:teuthology.orchestra.run.vm02.stdout:210453397532 2026-03-21T06:55:59.565 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.6 2026-03-21T06:55:59.684 INFO:teuthology.orchestra.run.vm02.stdout:240518168580 2026-03-21T06:55:59.684 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.7 2026-03-21T06:55:59.693 INFO:teuthology.orchestra.run.vm02.stdout:176093659167 2026-03-21T06:55:59.693 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.5 2026-03-21T06:55:59.745 INFO:teuthology.orchestra.run.vm02.stdout:77309411389 2026-03-21T06:55:59.745 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.1 2026-03-21T06:55:59.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: audit 2026-03-21T06:55:58.496779+0000 mon.a (mon.0) 785 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"}]': finished 2026-03-21T06:55:59.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: audit 2026-03-21T06:55:58.496779+0000 mon.a (mon.0) 785 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"}]': finished 2026-03-21T06:55:59.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: cluster 2026-03-21T06:55:58.506900+0000 mon.a (mon.0) 786 : cluster [DBG] osdmap e61: 8 total, 8 up, 8 in 2026-03-21T06:55:59.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: cluster 2026-03-21T06:55:58.506900+0000 mon.a (mon.0) 786 : cluster [DBG] osdmap e61: 8 total, 8 up, 8 in 2026-03-21T06:55:59.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: cluster 2026-03-21T06:55:58.507214+0000 mon.a (mon.0) 787 : cluster [DBG] mgrmap e15: x(active, since 7m) 2026-03-21T06:55:59.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: cluster 2026-03-21T06:55:58.507214+0000 mon.a (mon.0) 787 : cluster [DBG] mgrmap e15: x(active, since 7m) 2026-03-21T06:55:59.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: audit 2026-03-21T06:55:58.525571+0000 mon.a (mon.0) 788 : audit [DBG] from='client.? 192.168.123.102:0/1524200044' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:59.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: audit 2026-03-21T06:55:58.525571+0000 mon.a (mon.0) 788 : audit [DBG] from='client.? 192.168.123.102:0/1524200044' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:59.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: audit 2026-03-21T06:55:58.787402+0000 mon.a (mon.0) 789 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"} : dispatch 2026-03-21T06:55:59.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: audit 2026-03-21T06:55:58.787402+0000 mon.a (mon.0) 789 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"} : dispatch 2026-03-21T06:55:59.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: audit 2026-03-21T06:55:58.960524+0000 mon.a (mon.0) 790 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"}]': finished 2026-03-21T06:55:59.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: audit 2026-03-21T06:55:58.960524+0000 mon.a (mon.0) 790 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"}]': finished 2026-03-21T06:55:59.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: cluster 2026-03-21T06:55:58.963941+0000 mon.a (mon.0) 791 : cluster [DBG] osdmap e62: 8 total, 8 up, 8 in 2026-03-21T06:55:59.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: cluster 2026-03-21T06:55:58.963941+0000 mon.a (mon.0) 791 : cluster [DBG] osdmap e62: 8 total, 8 up, 8 in 2026-03-21T06:55:59.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: audit 2026-03-21T06:55:59.289194+0000 mon.a (mon.0) 792 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"} : dispatch 2026-03-21T06:55:59.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:55:59 vm04 bash[20194]: audit 2026-03-21T06:55:59.289194+0000 mon.a (mon.0) 792 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"} : dispatch 2026-03-21T06:55:59.769 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: audit 2026-03-21T06:55:58.496779+0000 mon.a (mon.0) 785 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"}]': finished 2026-03-21T06:55:59.769 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: audit 2026-03-21T06:55:58.496779+0000 mon.a (mon.0) 785 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"}]': finished 2026-03-21T06:55:59.769 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: cluster 2026-03-21T06:55:58.506900+0000 mon.a (mon.0) 786 : cluster [DBG] osdmap e61: 8 total, 8 up, 8 in 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: cluster 2026-03-21T06:55:58.506900+0000 mon.a (mon.0) 786 : cluster [DBG] osdmap e61: 8 total, 8 up, 8 in 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: cluster 2026-03-21T06:55:58.507214+0000 mon.a (mon.0) 787 : cluster [DBG] mgrmap e15: x(active, since 7m) 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: cluster 2026-03-21T06:55:58.507214+0000 mon.a (mon.0) 787 : cluster [DBG] mgrmap e15: x(active, since 7m) 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: audit 2026-03-21T06:55:58.525571+0000 mon.a (mon.0) 788 : audit [DBG] from='client.? 192.168.123.102:0/1524200044' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: audit 2026-03-21T06:55:58.525571+0000 mon.a (mon.0) 788 : audit [DBG] from='client.? 192.168.123.102:0/1524200044' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: audit 2026-03-21T06:55:58.787402+0000 mon.a (mon.0) 789 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"} : dispatch 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: audit 2026-03-21T06:55:58.787402+0000 mon.a (mon.0) 789 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"} : dispatch 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: audit 2026-03-21T06:55:58.960524+0000 mon.a (mon.0) 790 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"}]': finished 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: audit 2026-03-21T06:55:58.960524+0000 mon.a (mon.0) 790 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"}]': finished 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: cluster 2026-03-21T06:55:58.963941+0000 mon.a (mon.0) 791 : cluster [DBG] osdmap e62: 8 total, 8 up, 8 in 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: cluster 2026-03-21T06:55:58.963941+0000 mon.a (mon.0) 791 : cluster [DBG] osdmap e62: 8 total, 8 up, 8 in 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: audit 2026-03-21T06:55:59.289194+0000 mon.a (mon.0) 792 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"} : dispatch 2026-03-21T06:55:59.770 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[17657]: audit 2026-03-21T06:55:59.289194+0000 mon.a (mon.0) 792 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"} : dispatch 2026-03-21T06:55:59.796 INFO:teuthology.orchestra.run.vm02.stdout:51539607615 2026-03-21T06:55:59.796 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.0 2026-03-21T06:55:59.832 INFO:teuthology.orchestra.run.vm02.stdout:90194313277 2026-03-21T06:55:59.832 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.2 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: audit 2026-03-21T06:55:58.496779+0000 mon.a (mon.0) 785 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"}]': finished 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: audit 2026-03-21T06:55:58.496779+0000 mon.a (mon.0) 785 : audit [INF] from='client.? 192.168.123.102:0/3361689070' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/4266123507"}]': finished 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: cluster 2026-03-21T06:55:58.506900+0000 mon.a (mon.0) 786 : cluster [DBG] osdmap e61: 8 total, 8 up, 8 in 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: cluster 2026-03-21T06:55:58.506900+0000 mon.a (mon.0) 786 : cluster [DBG] osdmap e61: 8 total, 8 up, 8 in 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: cluster 2026-03-21T06:55:58.507214+0000 mon.a (mon.0) 787 : cluster [DBG] mgrmap e15: x(active, since 7m) 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: cluster 2026-03-21T06:55:58.507214+0000 mon.a (mon.0) 787 : cluster [DBG] mgrmap e15: x(active, since 7m) 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: audit 2026-03-21T06:55:58.525571+0000 mon.a (mon.0) 788 : audit [DBG] from='client.? 192.168.123.102:0/1524200044' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: audit 2026-03-21T06:55:58.525571+0000 mon.a (mon.0) 788 : audit [DBG] from='client.? 192.168.123.102:0/1524200044' entity='client.admin' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: audit 2026-03-21T06:55:58.787402+0000 mon.a (mon.0) 789 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"} : dispatch 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: audit 2026-03-21T06:55:58.787402+0000 mon.a (mon.0) 789 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"} : dispatch 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: audit 2026-03-21T06:55:58.960524+0000 mon.a (mon.0) 790 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"}]': finished 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: audit 2026-03-21T06:55:58.960524+0000 mon.a (mon.0) 790 : audit [INF] from='client.? 192.168.123.102:0/2240416376' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1909593542"}]': finished 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: cluster 2026-03-21T06:55:58.963941+0000 mon.a (mon.0) 791 : cluster [DBG] osdmap e62: 8 total, 8 up, 8 in 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: cluster 2026-03-21T06:55:58.963941+0000 mon.a (mon.0) 791 : cluster [DBG] osdmap e62: 8 total, 8 up, 8 in 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: audit 2026-03-21T06:55:59.289194+0000 mon.a (mon.0) 792 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"} : dispatch 2026-03-21T06:55:59.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:55:59 vm07 bash[19945]: audit 2026-03-21T06:55:59.289194+0000 mon.a (mon.0) 792 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"} : dispatch 2026-03-21T06:55:59.904 INFO:teuthology.orchestra.run.vm02.stdout:115964117050 2026-03-21T06:55:59.904 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.3 2026-03-21T06:55:59.914 INFO:teuthology.orchestra.run.vm02.stdout:137438953529 2026-03-21T06:55:59.915 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:55:59.915 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.4 2026-03-21T06:56:00.047 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[49158]: debug Successfully removed blocklist entry 2026-03-21T06:56:00.047 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:55:59 vm02 bash[49158]: debug Removing blocklisted entry for this host : 192.168.123.102:0/1580070482 2026-03-21T06:56:00.197 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:00.197 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:00.295 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:00.302 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:00.371 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:00.477 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:00.524 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:00.588 INFO:teuthology.orchestra.run.vm02.stdout:210453397532 2026-03-21T06:56:00.599 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:00 vm02 bash[17657]: cluster 2026-03-21T06:55:59.604675+0000 mgr.x (mgr.14152) 346 : cluster [DBG] pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:56:00.599 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:00 vm02 bash[17657]: cluster 2026-03-21T06:55:59.604675+0000 mgr.x (mgr.14152) 346 : cluster [DBG] pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:56:00.599 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:00 vm02 bash[17657]: audit 2026-03-21T06:55:59.963617+0000 mon.a (mon.0) 793 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"}]': finished 2026-03-21T06:56:00.599 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:00 vm02 bash[17657]: audit 2026-03-21T06:55:59.963617+0000 mon.a (mon.0) 793 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"}]': finished 2026-03-21T06:56:00.599 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:00 vm02 bash[17657]: cluster 2026-03-21T06:55:59.969380+0000 mon.a (mon.0) 794 : cluster [DBG] osdmap e63: 8 total, 8 up, 8 in 2026-03-21T06:56:00.599 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:00 vm02 bash[17657]: cluster 2026-03-21T06:55:59.969380+0000 mon.a (mon.0) 794 : cluster [DBG] osdmap e63: 8 total, 8 up, 8 in 2026-03-21T06:56:00.599 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:00 vm02 bash[17657]: audit 2026-03-21T06:56:00.393646+0000 mon.c (mon.1) 30 : audit [INF] from='client.? 192.168.123.102:0/2331242088' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:00.600 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:00 vm02 bash[17657]: audit 2026-03-21T06:56:00.393646+0000 mon.c (mon.1) 30 : audit [INF] from='client.? 192.168.123.102:0/2331242088' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:00.600 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:00 vm02 bash[17657]: audit 2026-03-21T06:56:00.395784+0000 mon.a (mon.0) 795 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:00.600 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:00 vm02 bash[17657]: audit 2026-03-21T06:56:00.395784+0000 mon.a (mon.0) 795 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:00.749 INFO:tasks.cephadm.ceph_manager.ceph:need seq 210453397532 got 210453397532 for osd.6 2026-03-21T06:56:00.749 DEBUG:teuthology.parallel:result is None 2026-03-21T06:56:00.837 INFO:teuthology.orchestra.run.vm02.stdout:176093659167 2026-03-21T06:56:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:00 vm07 bash[19945]: cluster 2026-03-21T06:55:59.604675+0000 mgr.x (mgr.14152) 346 : cluster [DBG] pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:56:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:00 vm07 bash[19945]: cluster 2026-03-21T06:55:59.604675+0000 mgr.x (mgr.14152) 346 : cluster [DBG] pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:56:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:00 vm07 bash[19945]: audit 2026-03-21T06:55:59.963617+0000 mon.a (mon.0) 793 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"}]': finished 2026-03-21T06:56:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:00 vm07 bash[19945]: audit 2026-03-21T06:55:59.963617+0000 mon.a (mon.0) 793 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"}]': finished 2026-03-21T06:56:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:00 vm07 bash[19945]: cluster 2026-03-21T06:55:59.969380+0000 mon.a (mon.0) 794 : cluster [DBG] osdmap e63: 8 total, 8 up, 8 in 2026-03-21T06:56:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:00 vm07 bash[19945]: cluster 2026-03-21T06:55:59.969380+0000 mon.a (mon.0) 794 : cluster [DBG] osdmap e63: 8 total, 8 up, 8 in 2026-03-21T06:56:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:00 vm07 bash[19945]: audit 2026-03-21T06:56:00.393646+0000 mon.c (mon.1) 30 : audit [INF] from='client.? 192.168.123.102:0/2331242088' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:00 vm07 bash[19945]: audit 2026-03-21T06:56:00.393646+0000 mon.c (mon.1) 30 : audit [INF] from='client.? 192.168.123.102:0/2331242088' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:00 vm07 bash[19945]: audit 2026-03-21T06:56:00.395784+0000 mon.a (mon.0) 795 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:00.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:00 vm07 bash[19945]: audit 2026-03-21T06:56:00.395784+0000 mon.a (mon.0) 795 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:00.913 INFO:teuthology.orchestra.run.vm02.stdout:77309411388 2026-03-21T06:56:00.934 INFO:teuthology.orchestra.run.vm02.stdout:240518168580 2026-03-21T06:56:00.982 INFO:tasks.cephadm.ceph_manager.ceph:need seq 176093659167 got 176093659167 for osd.5 2026-03-21T06:56:00.982 DEBUG:teuthology.parallel:result is None 2026-03-21T06:56:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:00 vm04 bash[20194]: cluster 2026-03-21T06:55:59.604675+0000 mgr.x (mgr.14152) 346 : cluster [DBG] pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:56:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:00 vm04 bash[20194]: cluster 2026-03-21T06:55:59.604675+0000 mgr.x (mgr.14152) 346 : cluster [DBG] pgmap v287: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 341 B/s wr, 0 op/s 2026-03-21T06:56:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:00 vm04 bash[20194]: audit 2026-03-21T06:55:59.963617+0000 mon.a (mon.0) 793 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"}]': finished 2026-03-21T06:56:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:00 vm04 bash[20194]: audit 2026-03-21T06:55:59.963617+0000 mon.a (mon.0) 793 : audit [INF] from='client.? 192.168.123.102:0/4026402835' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1252995422"}]': finished 2026-03-21T06:56:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:00 vm04 bash[20194]: cluster 2026-03-21T06:55:59.969380+0000 mon.a (mon.0) 794 : cluster [DBG] osdmap e63: 8 total, 8 up, 8 in 2026-03-21T06:56:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:00 vm04 bash[20194]: cluster 2026-03-21T06:55:59.969380+0000 mon.a (mon.0) 794 : cluster [DBG] osdmap e63: 8 total, 8 up, 8 in 2026-03-21T06:56:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:00 vm04 bash[20194]: audit 2026-03-21T06:56:00.393646+0000 mon.c (mon.1) 30 : audit [INF] from='client.? 192.168.123.102:0/2331242088' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:01.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:00 vm04 bash[20194]: audit 2026-03-21T06:56:00.393646+0000 mon.c (mon.1) 30 : audit [INF] from='client.? 192.168.123.102:0/2331242088' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:01.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:00 vm04 bash[20194]: audit 2026-03-21T06:56:00.395784+0000 mon.a (mon.0) 795 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:01.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:00 vm04 bash[20194]: audit 2026-03-21T06:56:00.395784+0000 mon.a (mon.0) 795 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"} : dispatch 2026-03-21T06:56:01.026 INFO:tasks.cephadm.ceph_manager.ceph:need seq 77309411389 got 77309411388 for osd.1 2026-03-21T06:56:01.045 INFO:teuthology.orchestra.run.vm02.stdout:115964117049 2026-03-21T06:56:01.074 INFO:teuthology.orchestra.run.vm02.stdout:90194313276 2026-03-21T06:56:01.083 INFO:tasks.cephadm.ceph_manager.ceph:need seq 240518168580 got 240518168580 for osd.7 2026-03-21T06:56:01.083 DEBUG:teuthology.parallel:result is None 2026-03-21T06:56:01.085 INFO:teuthology.orchestra.run.vm02.stdout:51539607615 2026-03-21T06:56:01.120 INFO:teuthology.orchestra.run.vm02.stdout:137438953528 2026-03-21T06:56:01.149 INFO:tasks.cephadm.ceph_manager.ceph:need seq 115964117050 got 115964117049 for osd.3 2026-03-21T06:56:01.174 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[49158]: debug Successfully removed blocklist entry 2026-03-21T06:56:01.174 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[49158]: debug Removing blocklisted entry for this host : 192.168.123.102:6801/260656842 2026-03-21T06:56:01.236 INFO:tasks.cephadm.ceph_manager.ceph:need seq 90194313277 got 90194313276 for osd.2 2026-03-21T06:56:01.244 INFO:tasks.cephadm.ceph_manager.ceph:need seq 137438953529 got 137438953528 for osd.4 2026-03-21T06:56:01.262 INFO:tasks.cephadm.ceph_manager.ceph:need seq 51539607615 got 51539607615 for osd.0 2026-03-21T06:56:01.262 DEBUG:teuthology.parallel:result is None 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:00.583635+0000 mon.b (mon.2) 19 : audit [DBG] from='client.? 192.168.123.102:0/2151639515' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 6} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:00.583635+0000 mon.b (mon.2) 19 : audit [DBG] from='client.? 192.168.123.102:0/2151639515' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 6} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:00.827859+0000 mon.b (mon.2) 20 : audit [DBG] from='client.? 192.168.123.102:0/4162555035' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 5} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:00.827859+0000 mon.b (mon.2) 20 : audit [DBG] from='client.? 192.168.123.102:0/4162555035' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 5} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:00.907805+0000 mon.b (mon.2) 21 : audit [DBG] from='client.? 192.168.123.102:0/275673320' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:00.907805+0000 mon.b (mon.2) 21 : audit [DBG] from='client.? 192.168.123.102:0/275673320' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:00.933535+0000 mon.a (mon.0) 796 : audit [DBG] from='client.? 192.168.123.102:0/1115182363' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 7} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:00.933535+0000 mon.a (mon.0) 796 : audit [DBG] from='client.? 192.168.123.102:0/1115182363' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 7} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:00.966416+0000 mon.a (mon.0) 797 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"}]': finished 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:00.966416+0000 mon.a (mon.0) 797 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"}]': finished 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: cluster 2026-03-21T06:56:00.968754+0000 mon.a (mon.0) 798 : cluster [DBG] osdmap e64: 8 total, 8 up, 8 in 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: cluster 2026-03-21T06:56:00.968754+0000 mon.a (mon.0) 798 : cluster [DBG] osdmap e64: 8 total, 8 up, 8 in 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.045770+0000 mon.c (mon.1) 31 : audit [DBG] from='client.? 192.168.123.102:0/4165751322' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.045770+0000 mon.c (mon.1) 31 : audit [DBG] from='client.? 192.168.123.102:0/4165751322' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.050926+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.050926+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.067326+0000 mon.a (mon.0) 800 : audit [DBG] from='client.? 192.168.123.102:0/2837601752' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.067326+0000 mon.a (mon.0) 800 : audit [DBG] from='client.? 192.168.123.102:0/2837601752' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.083367+0000 mon.c (mon.1) 32 : audit [DBG] from='client.? 192.168.123.102:0/3343516232' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.083367+0000 mon.c (mon.1) 32 : audit [DBG] from='client.? 192.168.123.102:0/3343516232' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.122039+0000 mon.a (mon.0) 801 : audit [DBG] from='client.? 192.168.123.102:0/679367926' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.122039+0000 mon.a (mon.0) 801 : audit [DBG] from='client.? 192.168.123.102:0/679367926' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.280928+0000 mon.a (mon.0) 802 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"} : dispatch 2026-03-21T06:56:01.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:01 vm02 bash[17657]: audit 2026-03-21T06:56:01.280928+0000 mon.a (mon.0) 802 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:00.583635+0000 mon.b (mon.2) 19 : audit [DBG] from='client.? 192.168.123.102:0/2151639515' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 6} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:00.583635+0000 mon.b (mon.2) 19 : audit [DBG] from='client.? 192.168.123.102:0/2151639515' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 6} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:00.827859+0000 mon.b (mon.2) 20 : audit [DBG] from='client.? 192.168.123.102:0/4162555035' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 5} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:00.827859+0000 mon.b (mon.2) 20 : audit [DBG] from='client.? 192.168.123.102:0/4162555035' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 5} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:00.907805+0000 mon.b (mon.2) 21 : audit [DBG] from='client.? 192.168.123.102:0/275673320' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:00.907805+0000 mon.b (mon.2) 21 : audit [DBG] from='client.? 192.168.123.102:0/275673320' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:00.933535+0000 mon.a (mon.0) 796 : audit [DBG] from='client.? 192.168.123.102:0/1115182363' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 7} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:00.933535+0000 mon.a (mon.0) 796 : audit [DBG] from='client.? 192.168.123.102:0/1115182363' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 7} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:00.966416+0000 mon.a (mon.0) 797 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"}]': finished 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:00.966416+0000 mon.a (mon.0) 797 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"}]': finished 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: cluster 2026-03-21T06:56:00.968754+0000 mon.a (mon.0) 798 : cluster [DBG] osdmap e64: 8 total, 8 up, 8 in 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: cluster 2026-03-21T06:56:00.968754+0000 mon.a (mon.0) 798 : cluster [DBG] osdmap e64: 8 total, 8 up, 8 in 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.045770+0000 mon.c (mon.1) 31 : audit [DBG] from='client.? 192.168.123.102:0/4165751322' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.045770+0000 mon.c (mon.1) 31 : audit [DBG] from='client.? 192.168.123.102:0/4165751322' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.050926+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.050926+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.067326+0000 mon.a (mon.0) 800 : audit [DBG] from='client.? 192.168.123.102:0/2837601752' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.067326+0000 mon.a (mon.0) 800 : audit [DBG] from='client.? 192.168.123.102:0/2837601752' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.083367+0000 mon.c (mon.1) 32 : audit [DBG] from='client.? 192.168.123.102:0/3343516232' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.083367+0000 mon.c (mon.1) 32 : audit [DBG] from='client.? 192.168.123.102:0/3343516232' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.122039+0000 mon.a (mon.0) 801 : audit [DBG] from='client.? 192.168.123.102:0/679367926' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.122039+0000 mon.a (mon.0) 801 : audit [DBG] from='client.? 192.168.123.102:0/679367926' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.280928+0000 mon.a (mon.0) 802 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"} : dispatch 2026-03-21T06:56:01.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:01 vm07 bash[19945]: audit 2026-03-21T06:56:01.280928+0000 mon.a (mon.0) 802 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"} : dispatch 2026-03-21T06:56:02.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:00.583635+0000 mon.b (mon.2) 19 : audit [DBG] from='client.? 192.168.123.102:0/2151639515' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 6} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:00.583635+0000 mon.b (mon.2) 19 : audit [DBG] from='client.? 192.168.123.102:0/2151639515' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 6} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:00.827859+0000 mon.b (mon.2) 20 : audit [DBG] from='client.? 192.168.123.102:0/4162555035' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 5} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:00.827859+0000 mon.b (mon.2) 20 : audit [DBG] from='client.? 192.168.123.102:0/4162555035' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 5} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:00.907805+0000 mon.b (mon.2) 21 : audit [DBG] from='client.? 192.168.123.102:0/275673320' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:00.907805+0000 mon.b (mon.2) 21 : audit [DBG] from='client.? 192.168.123.102:0/275673320' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:00.933535+0000 mon.a (mon.0) 796 : audit [DBG] from='client.? 192.168.123.102:0/1115182363' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 7} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:00.933535+0000 mon.a (mon.0) 796 : audit [DBG] from='client.? 192.168.123.102:0/1115182363' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 7} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:00.966416+0000 mon.a (mon.0) 797 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"}]': finished 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:00.966416+0000 mon.a (mon.0) 797 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1580070482"}]': finished 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: cluster 2026-03-21T06:56:00.968754+0000 mon.a (mon.0) 798 : cluster [DBG] osdmap e64: 8 total, 8 up, 8 in 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: cluster 2026-03-21T06:56:00.968754+0000 mon.a (mon.0) 798 : cluster [DBG] osdmap e64: 8 total, 8 up, 8 in 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.045770+0000 mon.c (mon.1) 31 : audit [DBG] from='client.? 192.168.123.102:0/4165751322' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.045770+0000 mon.c (mon.1) 31 : audit [DBG] from='client.? 192.168.123.102:0/4165751322' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.050926+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.050926+0000 mon.a (mon.0) 799 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.067326+0000 mon.a (mon.0) 800 : audit [DBG] from='client.? 192.168.123.102:0/2837601752' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.067326+0000 mon.a (mon.0) 800 : audit [DBG] from='client.? 192.168.123.102:0/2837601752' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.083367+0000 mon.c (mon.1) 32 : audit [DBG] from='client.? 192.168.123.102:0/3343516232' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.083367+0000 mon.c (mon.1) 32 : audit [DBG] from='client.? 192.168.123.102:0/3343516232' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 0} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.122039+0000 mon.a (mon.0) 801 : audit [DBG] from='client.? 192.168.123.102:0/679367926' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.122039+0000 mon.a (mon.0) 801 : audit [DBG] from='client.? 192.168.123.102:0/679367926' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.280928+0000 mon.a (mon.0) 802 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"} : dispatch 2026-03-21T06:56:02.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:01 vm04 bash[20194]: audit 2026-03-21T06:56:01.280928+0000 mon.a (mon.0) 802 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"} : dispatch 2026-03-21T06:56:02.027 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.1 2026-03-21T06:56:02.150 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.3 2026-03-21T06:56:02.238 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.2 2026-03-21T06:56:02.245 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph osd last-stat-seq osd.4 2026-03-21T06:56:02.305 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:02.310 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[49158]: debug Successfully removed blocklist entry 2026-03-21T06:56:02.310 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[49158]: debug Removing blocklisted entry for this host : 192.168.123.102:6800/260656842 2026-03-21T06:56:02.447 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:02.577 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:02.594 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:02.624 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: cluster 2026-03-21T06:56:01.605005+0000 mgr.x (mgr.14152) 347 : cluster [DBG] pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:02.624 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: cluster 2026-03-21T06:56:01.605005+0000 mgr.x (mgr.14152) 347 : cluster [DBG] pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:02.624 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: audit 2026-03-21T06:56:02.052223+0000 mon.a (mon.0) 803 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"}]': finished 2026-03-21T06:56:02.624 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: audit 2026-03-21T06:56:02.052223+0000 mon.a (mon.0) 803 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"}]': finished 2026-03-21T06:56:02.625 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: cluster 2026-03-21T06:56:02.055437+0000 mon.a (mon.0) 804 : cluster [DBG] osdmap e65: 8 total, 8 up, 8 in 2026-03-21T06:56:02.625 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: cluster 2026-03-21T06:56:02.055437+0000 mon.a (mon.0) 804 : cluster [DBG] osdmap e65: 8 total, 8 up, 8 in 2026-03-21T06:56:02.625 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: audit 2026-03-21T06:56:02.297994+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:02.625 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: audit 2026-03-21T06:56:02.297994+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:02.625 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: audit 2026-03-21T06:56:02.298591+0000 mon.c (mon.1) 33 : audit [INF] from='client.? 192.168.123.102:0/2970604024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:02.625 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: audit 2026-03-21T06:56:02.298591+0000 mon.c (mon.1) 33 : audit [INF] from='client.? 192.168.123.102:0/2970604024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:02.625 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: audit 2026-03-21T06:56:02.300854+0000 mon.a (mon.0) 806 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:02.625 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: audit 2026-03-21T06:56:02.300854+0000 mon.a (mon.0) 806 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:02.625 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: audit 2026-03-21T06:56:02.303346+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:02.625 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:02 vm02 bash[17657]: audit 2026-03-21T06:56:02.303346+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:02.757 INFO:teuthology.orchestra.run.vm02.stdout:77309411389 2026-03-21T06:56:02.840 INFO:tasks.cephadm.ceph_manager.ceph:need seq 77309411389 got 77309411389 for osd.1 2026-03-21T06:56:02.841 DEBUG:teuthology.parallel:result is None 2026-03-21T06:56:02.893 INFO:teuthology.orchestra.run.vm02.stdout:115964117051 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: cluster 2026-03-21T06:56:01.605005+0000 mgr.x (mgr.14152) 347 : cluster [DBG] pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: cluster 2026-03-21T06:56:01.605005+0000 mgr.x (mgr.14152) 347 : cluster [DBG] pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: audit 2026-03-21T06:56:02.052223+0000 mon.a (mon.0) 803 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"}]': finished 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: audit 2026-03-21T06:56:02.052223+0000 mon.a (mon.0) 803 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"}]': finished 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: cluster 2026-03-21T06:56:02.055437+0000 mon.a (mon.0) 804 : cluster [DBG] osdmap e65: 8 total, 8 up, 8 in 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: cluster 2026-03-21T06:56:02.055437+0000 mon.a (mon.0) 804 : cluster [DBG] osdmap e65: 8 total, 8 up, 8 in 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: audit 2026-03-21T06:56:02.297994+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: audit 2026-03-21T06:56:02.297994+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: audit 2026-03-21T06:56:02.298591+0000 mon.c (mon.1) 33 : audit [INF] from='client.? 192.168.123.102:0/2970604024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: audit 2026-03-21T06:56:02.298591+0000 mon.c (mon.1) 33 : audit [INF] from='client.? 192.168.123.102:0/2970604024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: audit 2026-03-21T06:56:02.300854+0000 mon.a (mon.0) 806 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: audit 2026-03-21T06:56:02.300854+0000 mon.a (mon.0) 806 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: audit 2026-03-21T06:56:02.303346+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:02.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:02 vm07 bash[19945]: audit 2026-03-21T06:56:02.303346+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:02.971 INFO:tasks.cephadm.ceph_manager.ceph:need seq 115964117050 got 115964117051 for osd.3 2026-03-21T06:56:02.972 INFO:teuthology.orchestra.run.vm02.stdout:90194313277 2026-03-21T06:56:02.972 DEBUG:teuthology.parallel:result is None 2026-03-21T06:56:03.004 INFO:teuthology.orchestra.run.vm02.stdout:137438953529 2026-03-21T06:56:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: cluster 2026-03-21T06:56:01.605005+0000 mgr.x (mgr.14152) 347 : cluster [DBG] pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: cluster 2026-03-21T06:56:01.605005+0000 mgr.x (mgr.14152) 347 : cluster [DBG] pgmap v290: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: audit 2026-03-21T06:56:02.052223+0000 mon.a (mon.0) 803 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"}]': finished 2026-03-21T06:56:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: audit 2026-03-21T06:56:02.052223+0000 mon.a (mon.0) 803 : audit [INF] from='client.? 192.168.123.102:0/784798024' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6801/260656842"}]': finished 2026-03-21T06:56:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: cluster 2026-03-21T06:56:02.055437+0000 mon.a (mon.0) 804 : cluster [DBG] osdmap e65: 8 total, 8 up, 8 in 2026-03-21T06:56:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: cluster 2026-03-21T06:56:02.055437+0000 mon.a (mon.0) 804 : cluster [DBG] osdmap e65: 8 total, 8 up, 8 in 2026-03-21T06:56:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: audit 2026-03-21T06:56:02.297994+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: audit 2026-03-21T06:56:02.297994+0000 mon.a (mon.0) 805 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: audit 2026-03-21T06:56:02.298591+0000 mon.c (mon.1) 33 : audit [INF] from='client.? 192.168.123.102:0/2970604024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: audit 2026-03-21T06:56:02.298591+0000 mon.c (mon.1) 33 : audit [INF] from='client.? 192.168.123.102:0/2970604024' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: audit 2026-03-21T06:56:02.300854+0000 mon.a (mon.0) 806 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: audit 2026-03-21T06:56:02.300854+0000 mon.a (mon.0) 806 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"} : dispatch 2026-03-21T06:56:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: audit 2026-03-21T06:56:02.303346+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:02 vm04 bash[20194]: audit 2026-03-21T06:56:02.303346+0000 mon.a (mon.0) 807 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.028 INFO:tasks.cephadm.ceph_manager.ceph:need seq 90194313277 got 90194313277 for osd.2 2026-03-21T06:56:03.028 DEBUG:teuthology.parallel:result is None 2026-03-21T06:56:03.086 INFO:tasks.cephadm.ceph_manager.ceph:need seq 137438953529 got 137438953529 for osd.4 2026-03-21T06:56:03.086 DEBUG:teuthology.parallel:result is None 2026-03-21T06:56:03.086 INFO:tasks.cephadm.ceph_manager.ceph:waiting for clean 2026-03-21T06:56:03.086 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph pg dump --format=json 2026-03-21T06:56:03.347 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:03.370 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[49158]: debug Successfully removed blocklist entry 2026-03-21T06:56:03.371 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[49158]: debug Removing blocklisted entry for this host : 192.168.123.102:0/2023406208 2026-03-21T06:56:03.705 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-21T06:56:03.705 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:56:03.762 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":293,"stamp":"2026-03-21T06:56:03.605373+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459688,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":59,"num_read_kb":44,"num_write":63,"num_write_kb":587,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":41,"ondisk_log_size":41,"up":12,"acting":12,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":12,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":6,"kb":167739392,"kb_used":223468,"kb_used_data":6332,"kb_used_omap":180,"kb_used_meta":216651,"kb_avail":167515924,"statfs":{"total":171765137408,"available":171536306176,"internally_reserved":0,"allocated":6483968,"data_stored":3520017,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":184925,"internal_metadata":221851043},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":13,"num_read_kb":7,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.000891"},"pg_stats":[{"pgid":"2.2","version":"59'2","reported_seq":30,"reported_epoch":63,"state":"active+clean","last_fresh":"2026-03-21T06:55:59.962255+0000","last_change":"2026-03-21T06:55:52.865584+0000","last_active":"2026-03-21T06:55:59.962255+0000","last_peered":"2026-03-21T06:55:59.962255+0000","last_clean":"2026-03-21T06:55:59.962255+0000","last_became_active":"2026-03-21T06:55:50.843936+0000","last_became_peered":"2026-03-21T06:55:50.843936+0000","last_unstale":"2026-03-21T06:55:59.962255+0000","last_undegraded":"2026-03-21T06:55:59.962255+0000","last_fullsized":"2026-03-21T06:55:59.962255+0000","mapping_epoch":56,"log_start":"0'0","ondisk_log_start":"0'0","created":56,"last_epoch_clean":57,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_clean_scrub_stamp":"2026-03-21T06:55:49.834055+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T12:25:02.426060+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00078937300000000001,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,2],"acting":[3,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"2.1","version":"57'1","reported_seq":27,"reported_epoch":62,"state":"active+clean","last_fresh":"2026-03-21T06:55:58.959560+0000","last_change":"2026-03-21T06:55:52.862697+0000","last_active":"2026-03-21T06:55:58.959560+0000","last_peered":"2026-03-21T06:55:58.959560+0000","last_clean":"2026-03-21T06:55:58.959560+0000","last_became_active":"2026-03-21T06:55:51.191297+0000","last_became_peered":"2026-03-21T06:55:51.191297+0000","last_unstale":"2026-03-21T06:55:58.959560+0000","last_undegraded":"2026-03-21T06:55:58.959560+0000","last_fullsized":"2026-03-21T06:55:58.959560+0000","mapping_epoch":56,"log_start":"0'0","ondisk_log_start":"0'0","created":56,"last_epoch_clean":57,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_clean_scrub_stamp":"2026-03-21T06:55:49.834055+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T13:31:04.917183+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00030340400000000001,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,1,0],"acting":[2,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"2.0","version":"59'6","reported_seq":45,"reported_epoch":63,"state":"active+clean","last_fresh":"2026-03-21T06:55:59.962287+0000","last_change":"2026-03-21T06:55:52.865114+0000","last_active":"2026-03-21T06:55:59.962287+0000","last_peered":"2026-03-21T06:55:59.962287+0000","last_clean":"2026-03-21T06:55:59.962287+0000","last_became_active":"2026-03-21T06:55:50.889602+0000","last_became_peered":"2026-03-21T06:55:50.889602+0000","last_unstale":"2026-03-21T06:55:59.962287+0000","last_undegraded":"2026-03-21T06:55:59.962287+0000","last_fullsized":"2026-03-21T06:55:59.962287+0000","mapping_epoch":56,"log_start":"0'0","ondisk_log_start":"0'0","created":56,"last_epoch_clean":57,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_clean_scrub_stamp":"2026-03-21T06:55:49.834055+0000","objects_scrubbed":0,"log_size":6,"log_dups_size":0,"ondisk_log_size":6,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T18:39:46.612484+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000312059,"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":13,"num_read_kb":7,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,6],"acting":[3,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"1.0","version":"23'32","reported_seq":107,"reported_epoch":63,"state":"active+clean","last_fresh":"2026-03-21T06:55:59.962255+0000","last_change":"2026-03-21T06:53:31.019655+0000","last_active":"2026-03-21T06:55:59.962255+0000","last_peered":"2026-03-21T06:55:59.962255+0000","last_clean":"2026-03-21T06:55:59.962255+0000","last_became_active":"2026-03-21T06:53:31.019201+0000","last_became_peered":"2026-03-21T06:53:31.019201+0000","last_unstale":"2026-03-21T06:55:59.962255+0000","last_undegraded":"2026-03-21T06:55:59.962255+0000","last_fullsized":"2026-03-21T06:55:59.962255+0000","mapping_epoch":41,"log_start":"0'0","ondisk_log_start":"0'0","created":22,"last_epoch_clean":42,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-21T06:51:08.627047+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-21T06:51:08.627047+0000","last_clean_scrub_stamp":"2026-03-21T06:51:08.627047+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T08:43:26.490403+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,2],"acting":[3,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":2,"num_pg":3,"stat_sum":{"num_bytes":408,"num_objects":3,"num_object_clones":0,"num_object_copies":9,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":3,"num_whiteouts":0,"num_read":13,"num_read_kb":7,"num_write":6,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1224,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":9,"ondisk_log_size":9,"up":9,"acting":9,"num_store_stats":6},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2314240,"data_stored":2296400,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":5}],"osd_stats":[{"osd":7,"up_from":56,"seq":240518168580,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27904,"kb_used_data":592,"kb_used_omap":8,"kb_used_meta":27255,"kb_avail":20939520,"statfs":{"total":21470642176,"available":21442068480,"internally_reserved":0,"allocated":606208,"data_stored":254436,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8296,"internal_metadata":27910040},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":49,"seq":210453397533,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27908,"kb_used_data":620,"kb_used_omap":13,"kb_used_meta":27250,"kb_avail":20939516,"statfs":{"total":21470642176,"available":21442064384,"internally_reserved":0,"allocated":634880,"data_stored":266616,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":13591,"internal_metadata":27904745},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":41,"seq":176093659168,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":28440,"kb_used_data":1072,"kb_used_omap":14,"kb_used_meta":27313,"kb_avail":20938984,"statfs":{"total":21470642176,"available":21441519616,"internally_reserved":0,"allocated":1097728,"data_stored":725896,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":14896,"internal_metadata":27968976},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":32,"seq":137438953530,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27516,"kb_used_data":652,"kb_used_omap":23,"kb_used_meta":26792,"kb_avail":20939908,"statfs":{"total":21470642176,"available":21442465792,"internally_reserved":0,"allocated":667648,"data_stored":283465,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":24483,"internal_metadata":27435101},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":27,"seq":115964117051,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27936,"kb_used_data":1088,"kb_used_omap":25,"kb_used_meta":26790,"kb_avail":20939488,"statfs":{"total":21470642176,"available":21442035712,"internally_reserved":0,"allocated":1114112,"data_stored":731704,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":25783,"internal_metadata":27433801},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":21,"seq":90194313277,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27920,"kb_used_data":1072,"kb_used_omap":28,"kb_used_meta":26787,"kb_avail":20939504,"statfs":{"total":21470642176,"available":21442052096,"internally_reserved":0,"allocated":1097728,"data_stored":725526,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":29691,"internal_metadata":27429893},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":18,"seq":77309411389,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27904,"kb_used_data":608,"kb_used_omap":31,"kb_used_meta":27232,"kb_avail":20939520,"statfs":{"total":21470642176,"available":21442068480,"internally_reserved":0,"allocated":622592,"data_stored":260358,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":32418,"internal_metadata":27885918},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":12,"seq":51539607616,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27940,"kb_used_data":628,"kb_used_omap":34,"kb_used_meta":27229,"kb_avail":20939484,"statfs":{"total":21470642176,"available":21442031616,"internally_reserved":0,"allocated":643072,"data_stored":272016,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":35767,"internal_metadata":27882569},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":408,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-21T06:56:03.762 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph pg dump --format=json 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:02.598844+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:02.598844+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:02.606770+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:02.606770+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:02.758549+0000 mon.a (mon.0) 810 : audit [DBG] from='client.? 192.168.123.102:0/233227809' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:02.758549+0000 mon.a (mon.0) 810 : audit [DBG] from='client.? 192.168.123.102:0/233227809' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:02.891823+0000 mon.a (mon.0) 811 : audit [DBG] from='client.? 192.168.123.102:0/4275780479' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:02.891823+0000 mon.a (mon.0) 811 : audit [DBG] from='client.? 192.168.123.102:0/4275780479' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:02.971244+0000 mon.c (mon.1) 34 : audit [DBG] from='client.? 192.168.123.102:0/4068279805' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:02.971244+0000 mon.c (mon.1) 34 : audit [DBG] from='client.? 192.168.123.102:0/4068279805' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.005121+0000 mon.a (mon.0) 812 : audit [DBG] from='client.? 192.168.123.102:0/599441177' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.005121+0000 mon.a (mon.0) 812 : audit [DBG] from='client.? 192.168.123.102:0/599441177' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.299098+0000 mon.a (mon.0) 813 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"}]': finished 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.299098+0000 mon.a (mon.0) 813 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"}]': finished 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: cluster 2026-03-21T06:56:03.301908+0000 mon.a (mon.0) 814 : cluster [DBG] osdmap e66: 8 total, 8 up, 8 in 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: cluster 2026-03-21T06:56:03.301908+0000 mon.a (mon.0) 814 : cluster [DBG] osdmap e66: 8 total, 8 up, 8 in 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.385767+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.385767+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.393164+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.393164+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.398782+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.398782+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.404793+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.404793+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.405881+0000 mon.a (mon.0) 819 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.405881+0000 mon.a (mon.0) 819 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.407916+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.407916+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.411941+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.411941+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.412855+0000 mon.a (mon.0) 822 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.412855+0000 mon.a (mon.0) 822 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.437470+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.437470+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.442267+0000 mon.a (mon.0) 824 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.442267+0000 mon.a (mon.0) 824 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.446638+0000 mon.a (mon.0) 825 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.446638+0000 mon.a (mon.0) 825 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.450969+0000 mon.a (mon.0) 826 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.450969+0000 mon.a (mon.0) 826 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.452659+0000 mon.a (mon.0) 827 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-list"} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.452659+0000 mon.a (mon.0) 827 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-list"} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.454133+0000 mon.a (mon.0) 828 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.454133+0000 mon.a (mon.0) 828 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"} : dispatch 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.457292+0000 mon.a (mon.0) 829 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.457292+0000 mon.a (mon.0) 829 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.890 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.462496+0000 mon.a (mon.0) 830 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm02"} : dispatch 2026-03-21T06:56:03.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.462496+0000 mon.a (mon.0) 830 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm02"} : dispatch 2026-03-21T06:56:03.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.468470+0000 mon.a (mon.0) 831 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.468470+0000 mon.a (mon.0) 831 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.469473+0000 mon.a (mon.0) 832 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm07"} : dispatch 2026-03-21T06:56:03.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.469473+0000 mon.a (mon.0) 832 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm07"} : dispatch 2026-03-21T06:56:03.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.474369+0000 mon.a (mon.0) 833 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.474369+0000 mon.a (mon.0) 833 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.475370+0000 mon.a (mon.0) 834 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:56:03.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.475370+0000 mon.a (mon.0) 834 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:56:03.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.559415+0000 mon.a (mon.0) 835 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"} : dispatch 2026-03-21T06:56:03.891 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:03 vm02 bash[17657]: audit 2026-03-21T06:56:03.559415+0000 mon.a (mon.0) 835 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"} : dispatch 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:02.598844+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:02.598844+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:02.606770+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:02.606770+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:02.758549+0000 mon.a (mon.0) 810 : audit [DBG] from='client.? 192.168.123.102:0/233227809' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:02.758549+0000 mon.a (mon.0) 810 : audit [DBG] from='client.? 192.168.123.102:0/233227809' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:02.891823+0000 mon.a (mon.0) 811 : audit [DBG] from='client.? 192.168.123.102:0/4275780479' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:02.891823+0000 mon.a (mon.0) 811 : audit [DBG] from='client.? 192.168.123.102:0/4275780479' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:02.971244+0000 mon.c (mon.1) 34 : audit [DBG] from='client.? 192.168.123.102:0/4068279805' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:02.971244+0000 mon.c (mon.1) 34 : audit [DBG] from='client.? 192.168.123.102:0/4068279805' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.005121+0000 mon.a (mon.0) 812 : audit [DBG] from='client.? 192.168.123.102:0/599441177' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.005121+0000 mon.a (mon.0) 812 : audit [DBG] from='client.? 192.168.123.102:0/599441177' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.299098+0000 mon.a (mon.0) 813 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"}]': finished 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.299098+0000 mon.a (mon.0) 813 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"}]': finished 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: cluster 2026-03-21T06:56:03.301908+0000 mon.a (mon.0) 814 : cluster [DBG] osdmap e66: 8 total, 8 up, 8 in 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: cluster 2026-03-21T06:56:03.301908+0000 mon.a (mon.0) 814 : cluster [DBG] osdmap e66: 8 total, 8 up, 8 in 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.385767+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.385767+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.393164+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.393164+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.398782+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.398782+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.404793+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.404793+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.405881+0000 mon.a (mon.0) 819 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.405881+0000 mon.a (mon.0) 819 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.407916+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.407916+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.411941+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.411941+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.412855+0000 mon.a (mon.0) 822 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.412855+0000 mon.a (mon.0) 822 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.437470+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.437470+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.442267+0000 mon.a (mon.0) 824 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.442267+0000 mon.a (mon.0) 824 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.446638+0000 mon.a (mon.0) 825 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.446638+0000 mon.a (mon.0) 825 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.450969+0000 mon.a (mon.0) 826 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.450969+0000 mon.a (mon.0) 826 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.452659+0000 mon.a (mon.0) 827 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-list"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.452659+0000 mon.a (mon.0) 827 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-list"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.454133+0000 mon.a (mon.0) 828 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.454133+0000 mon.a (mon.0) 828 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.457292+0000 mon.a (mon.0) 829 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.457292+0000 mon.a (mon.0) 829 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.462496+0000 mon.a (mon.0) 830 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm02"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.462496+0000 mon.a (mon.0) 830 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm02"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.468470+0000 mon.a (mon.0) 831 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.468470+0000 mon.a (mon.0) 831 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.469473+0000 mon.a (mon.0) 832 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm07"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.469473+0000 mon.a (mon.0) 832 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm07"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.474369+0000 mon.a (mon.0) 833 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.474369+0000 mon.a (mon.0) 833 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.475370+0000 mon.a (mon.0) 834 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.475370+0000 mon.a (mon.0) 834 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.559415+0000 mon.a (mon.0) 835 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"} : dispatch 2026-03-21T06:56:03.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:03 vm07 bash[19945]: audit 2026-03-21T06:56:03.559415+0000 mon.a (mon.0) 835 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"} : dispatch 2026-03-21T06:56:04.003 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:02.598844+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:02.598844+0000 mon.a (mon.0) 808 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:02.606770+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:02.606770+0000 mon.a (mon.0) 809 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:02.758549+0000 mon.a (mon.0) 810 : audit [DBG] from='client.? 192.168.123.102:0/233227809' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:02.758549+0000 mon.a (mon.0) 810 : audit [DBG] from='client.? 192.168.123.102:0/233227809' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 1} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:02.891823+0000 mon.a (mon.0) 811 : audit [DBG] from='client.? 192.168.123.102:0/4275780479' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:02.891823+0000 mon.a (mon.0) 811 : audit [DBG] from='client.? 192.168.123.102:0/4275780479' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 3} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:02.971244+0000 mon.c (mon.1) 34 : audit [DBG] from='client.? 192.168.123.102:0/4068279805' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:02.971244+0000 mon.c (mon.1) 34 : audit [DBG] from='client.? 192.168.123.102:0/4068279805' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 2} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.005121+0000 mon.a (mon.0) 812 : audit [DBG] from='client.? 192.168.123.102:0/599441177' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.005121+0000 mon.a (mon.0) 812 : audit [DBG] from='client.? 192.168.123.102:0/599441177' entity='client.admin' cmd={"prefix": "osd last-stat-seq", "id": 4} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.299098+0000 mon.a (mon.0) 813 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"}]': finished 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.299098+0000 mon.a (mon.0) 813 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:6800/260656842"}]': finished 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: cluster 2026-03-21T06:56:03.301908+0000 mon.a (mon.0) 814 : cluster [DBG] osdmap e66: 8 total, 8 up, 8 in 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: cluster 2026-03-21T06:56:03.301908+0000 mon.a (mon.0) 814 : cluster [DBG] osdmap e66: 8 total, 8 up, 8 in 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.385767+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.385767+0000 mon.a (mon.0) 815 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.393164+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.393164+0000 mon.a (mon.0) 816 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.398782+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.398782+0000 mon.a (mon.0) 817 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.404793+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.404793+0000 mon.a (mon.0) 818 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.405881+0000 mon.a (mon.0) 819 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.405881+0000 mon.a (mon.0) 819 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.407916+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.407916+0000 mon.a (mon.0) 820 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.411941+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.411941+0000 mon.a (mon.0) 821 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.412855+0000 mon.a (mon.0) 822 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.412855+0000 mon.a (mon.0) 822 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.437470+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.437470+0000 mon.a (mon.0) 823 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.442267+0000 mon.a (mon.0) 824 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.442267+0000 mon.a (mon.0) 824 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.446638+0000 mon.a (mon.0) 825 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.446638+0000 mon.a (mon.0) 825 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.450969+0000 mon.a (mon.0) 826 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.450969+0000 mon.a (mon.0) 826 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.452659+0000 mon.a (mon.0) 827 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-list"} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.452659+0000 mon.a (mon.0) 827 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-list"} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.454133+0000 mon.a (mon.0) 828 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"} : dispatch 2026-03-21T06:56:04.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.454133+0000 mon.a (mon.0) 828 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"} : dispatch 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.457292+0000 mon.a (mon.0) 829 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.457292+0000 mon.a (mon.0) 829 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.462496+0000 mon.a (mon.0) 830 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm02"} : dispatch 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.462496+0000 mon.a (mon.0) 830 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm02"} : dispatch 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.468470+0000 mon.a (mon.0) 831 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.468470+0000 mon.a (mon.0) 831 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.469473+0000 mon.a (mon.0) 832 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm07"} : dispatch 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.469473+0000 mon.a (mon.0) 832 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "dashboard iscsi-gateway-add", "name": "vm07"} : dispatch 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.474369+0000 mon.a (mon.0) 833 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.474369+0000 mon.a (mon.0) 833 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.475370+0000 mon.a (mon.0) 834 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.475370+0000 mon.a (mon.0) 834 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.559415+0000 mon.a (mon.0) 835 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"} : dispatch 2026-03-21T06:56:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:03 vm04 bash[20194]: audit 2026-03-21T06:56:03.559415+0000 mon.a (mon.0) 835 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"} : dispatch 2026-03-21T06:56:04.339 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:56:04.339 INFO:teuthology.orchestra.run.vm02.stderr:dumped all 2026-03-21T06:56:04.394 INFO:teuthology.orchestra.run.vm02.stdout:{"pg_ready":true,"pg_map":{"version":293,"stamp":"2026-03-21T06:56:03.605373+0000","last_osdmap_epoch":0,"last_pg_scan":0,"pg_stats_sum":{"stat_sum":{"num_bytes":459688,"num_objects":5,"num_object_clones":0,"num_object_copies":15,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":5,"num_whiteouts":0,"num_read":59,"num_read_kb":44,"num_write":63,"num_write_kb":587,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":41,"ondisk_log_size":41,"up":12,"acting":12,"num_store_stats":0},"osd_stats_sum":{"up_from":0,"seq":0,"num_pgs":12,"num_osds":8,"num_per_pool_osds":8,"num_per_pool_omap_osds":6,"kb":167739392,"kb_used":223468,"kb_used_data":6332,"kb_used_omap":180,"kb_used_meta":216651,"kb_avail":167515924,"statfs":{"total":171765137408,"available":171536306176,"internally_reserved":0,"allocated":6483968,"data_stored":3520017,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":184925,"internal_metadata":221851043},"hb_peers":[],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[],"network_ping_times":[]},"pg_stats_delta":{"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":13,"num_read_kb":7,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":0,"ondisk_log_size":0,"up":0,"acting":0,"num_store_stats":0,"stamp_delta":"4.000891"},"pg_stats":[{"pgid":"2.2","version":"59'2","reported_seq":30,"reported_epoch":63,"state":"active+clean","last_fresh":"2026-03-21T06:55:59.962255+0000","last_change":"2026-03-21T06:55:52.865584+0000","last_active":"2026-03-21T06:55:59.962255+0000","last_peered":"2026-03-21T06:55:59.962255+0000","last_clean":"2026-03-21T06:55:59.962255+0000","last_became_active":"2026-03-21T06:55:50.843936+0000","last_became_peered":"2026-03-21T06:55:50.843936+0000","last_unstale":"2026-03-21T06:55:59.962255+0000","last_undegraded":"2026-03-21T06:55:59.962255+0000","last_fullsized":"2026-03-21T06:55:59.962255+0000","mapping_epoch":56,"log_start":"0'0","ondisk_log_start":"0'0","created":56,"last_epoch_clean":57,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_clean_scrub_stamp":"2026-03-21T06:55:49.834055+0000","objects_scrubbed":0,"log_size":2,"log_dups_size":0,"ondisk_log_size":2,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T12:25:02.426060+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00078937300000000001,"stat_sum":{"num_bytes":19,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":2,"num_write_kb":2,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,1,2],"acting":[3,1,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"2.1","version":"57'1","reported_seq":27,"reported_epoch":62,"state":"active+clean","last_fresh":"2026-03-21T06:55:58.959560+0000","last_change":"2026-03-21T06:55:52.862697+0000","last_active":"2026-03-21T06:55:58.959560+0000","last_peered":"2026-03-21T06:55:58.959560+0000","last_clean":"2026-03-21T06:55:58.959560+0000","last_became_active":"2026-03-21T06:55:51.191297+0000","last_became_peered":"2026-03-21T06:55:51.191297+0000","last_unstale":"2026-03-21T06:55:58.959560+0000","last_undegraded":"2026-03-21T06:55:58.959560+0000","last_fullsized":"2026-03-21T06:55:58.959560+0000","mapping_epoch":56,"log_start":"0'0","ondisk_log_start":"0'0","created":56,"last_epoch_clean":57,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_clean_scrub_stamp":"2026-03-21T06:55:49.834055+0000","objects_scrubbed":0,"log_size":1,"log_dups_size":0,"ondisk_log_size":1,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T13:31:04.917183+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.00030340400000000001,"stat_sum":{"num_bytes":0,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":0,"num_read_kb":0,"num_write":0,"num_write_kb":0,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[2,1,0],"acting":[2,1,0],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":2,"acting_primary":2,"purged_snaps":[]},{"pgid":"2.0","version":"59'6","reported_seq":45,"reported_epoch":63,"state":"active+clean","last_fresh":"2026-03-21T06:55:59.962287+0000","last_change":"2026-03-21T06:55:52.865114+0000","last_active":"2026-03-21T06:55:59.962287+0000","last_peered":"2026-03-21T06:55:59.962287+0000","last_clean":"2026-03-21T06:55:59.962287+0000","last_became_active":"2026-03-21T06:55:50.889602+0000","last_became_peered":"2026-03-21T06:55:50.889602+0000","last_unstale":"2026-03-21T06:55:59.962287+0000","last_undegraded":"2026-03-21T06:55:59.962287+0000","last_fullsized":"2026-03-21T06:55:59.962287+0000","mapping_epoch":56,"log_start":"0'0","ondisk_log_start":"0'0","created":56,"last_epoch_clean":57,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-21T06:55:49.834055+0000","last_clean_scrub_stamp":"2026-03-21T06:55:49.834055+0000","objects_scrubbed":0,"log_size":6,"log_dups_size":0,"ondisk_log_size":6,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T18:39:46.612484+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0.000312059,"stat_sum":{"num_bytes":389,"num_objects":1,"num_object_clones":0,"num_object_copies":3,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":1,"num_whiteouts":0,"num_read":13,"num_read_kb":7,"num_write":4,"num_write_kb":1,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,6],"acting":[3,5,6],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]},{"pgid":"1.0","version":"23'32","reported_seq":107,"reported_epoch":63,"state":"active+clean","last_fresh":"2026-03-21T06:55:59.962255+0000","last_change":"2026-03-21T06:53:31.019655+0000","last_active":"2026-03-21T06:55:59.962255+0000","last_peered":"2026-03-21T06:55:59.962255+0000","last_clean":"2026-03-21T06:55:59.962255+0000","last_became_active":"2026-03-21T06:53:31.019201+0000","last_became_peered":"2026-03-21T06:53:31.019201+0000","last_unstale":"2026-03-21T06:55:59.962255+0000","last_undegraded":"2026-03-21T06:55:59.962255+0000","last_fullsized":"2026-03-21T06:55:59.962255+0000","mapping_epoch":41,"log_start":"0'0","ondisk_log_start":"0'0","created":22,"last_epoch_clean":42,"parent":"0.0","parent_split_bits":0,"last_scrub":"0'0","last_scrub_stamp":"2026-03-21T06:51:08.627047+0000","last_deep_scrub":"0'0","last_deep_scrub_stamp":"2026-03-21T06:51:08.627047+0000","last_clean_scrub_stamp":"2026-03-21T06:51:08.627047+0000","objects_scrubbed":0,"log_size":32,"log_dups_size":0,"ondisk_log_size":32,"stats_invalid":false,"dirty_stats_invalid":false,"omap_stats_invalid":false,"hitset_stats_invalid":false,"hitset_bytes_stats_invalid":false,"pin_stats_invalid":false,"manifest_stats_invalid":false,"snaptrimq_len":0,"last_scrub_duration":0,"scrub_schedule":"periodic scrub scheduled @ 2026-03-22T08:43:26.490403+0000","scrub_duration":0,"objects_trimmed":0,"snaptrim_duration":0,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"up":[3,5,2],"acting":[3,5,2],"avail_no_missing":[],"object_location_counts":[],"blocked_by":[],"up_primary":3,"acting_primary":3,"purged_snaps":[]}],"pool_stats":[{"poolid":2,"num_pg":3,"stat_sum":{"num_bytes":408,"num_objects":3,"num_object_clones":0,"num_object_copies":9,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":3,"num_whiteouts":0,"num_read":13,"num_read_kb":7,"num_write":6,"num_write_kb":3,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":0,"num_bytes_recovered":0,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":24576,"data_stored":1224,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":9,"ondisk_log_size":9,"up":9,"acting":9,"num_store_stats":6},{"poolid":1,"num_pg":1,"stat_sum":{"num_bytes":459280,"num_objects":2,"num_object_clones":0,"num_object_copies":6,"num_objects_missing_on_primary":0,"num_objects_missing":0,"num_objects_degraded":0,"num_objects_misplaced":0,"num_objects_unfound":0,"num_objects_dirty":2,"num_whiteouts":0,"num_read":46,"num_read_kb":37,"num_write":57,"num_write_kb":584,"num_scrub_errors":0,"num_shallow_scrub_errors":0,"num_deep_scrub_errors":0,"num_objects_recovered":4,"num_bytes_recovered":918560,"num_keys_recovered":0,"num_objects_omap":0,"num_objects_hit_set_archive":0,"num_bytes_hit_set_archive":0,"num_flush":0,"num_flush_kb":0,"num_evict":0,"num_evict_kb":0,"num_promote":0,"num_flush_mode_high":0,"num_flush_mode_low":0,"num_evict_mode_some":0,"num_evict_mode_full":0,"num_objects_pinned":0,"num_legacy_snapsets":0,"num_large_omap_objects":0,"num_objects_manifest":0,"num_omap_bytes":0,"num_omap_keys":0,"num_objects_repaired":0},"store_stats":{"total":0,"available":0,"internally_reserved":0,"allocated":2314240,"data_stored":2296400,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},"log_size":32,"ondisk_log_size":32,"up":3,"acting":3,"num_store_stats":5}],"osd_stats":[{"osd":7,"up_from":56,"seq":240518168580,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27904,"kb_used_data":592,"kb_used_omap":8,"kb_used_meta":27255,"kb_avail":20939520,"statfs":{"total":21470642176,"available":21442068480,"internally_reserved":0,"allocated":606208,"data_stored":254436,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":8296,"internal_metadata":27910040},"hb_peers":[0,1,2,3,4,5,6],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":6,"up_from":49,"seq":210453397533,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27908,"kb_used_data":620,"kb_used_omap":13,"kb_used_meta":27250,"kb_avail":20939516,"statfs":{"total":21470642176,"available":21442064384,"internally_reserved":0,"allocated":634880,"data_stored":266616,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":13591,"internal_metadata":27904745},"hb_peers":[0,1,2,3,4,5,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":5,"up_from":41,"seq":176093659168,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":28440,"kb_used_data":1072,"kb_used_omap":14,"kb_used_meta":27313,"kb_avail":20938984,"statfs":{"total":21470642176,"available":21441519616,"internally_reserved":0,"allocated":1097728,"data_stored":725896,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":14896,"internal_metadata":27968976},"hb_peers":[0,1,2,3,4,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":4,"up_from":32,"seq":137438953530,"num_pgs":0,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":0,"kb":20967424,"kb_used":27516,"kb_used_data":652,"kb_used_omap":23,"kb_used_meta":26792,"kb_avail":20939908,"statfs":{"total":21470642176,"available":21442465792,"internally_reserved":0,"allocated":667648,"data_stored":283465,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":24483,"internal_metadata":27435101},"hb_peers":[0,1,2,3,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":3,"up_from":27,"seq":115964117051,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27936,"kb_used_data":1088,"kb_used_omap":25,"kb_used_meta":26790,"kb_avail":20939488,"statfs":{"total":21470642176,"available":21442035712,"internally_reserved":0,"allocated":1114112,"data_stored":731704,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":25783,"internal_metadata":27433801},"hb_peers":[0,1,2,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":2,"up_from":21,"seq":90194313277,"num_pgs":3,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27920,"kb_used_data":1072,"kb_used_omap":28,"kb_used_meta":26787,"kb_avail":20939504,"statfs":{"total":21470642176,"available":21442052096,"internally_reserved":0,"allocated":1097728,"data_stored":725526,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":29691,"internal_metadata":27429893},"hb_peers":[0,1,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":1,"up_from":18,"seq":77309411389,"num_pgs":2,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27904,"kb_used_data":608,"kb_used_omap":31,"kb_used_meta":27232,"kb_avail":20939520,"statfs":{"total":21470642176,"available":21442068480,"internally_reserved":0,"allocated":622592,"data_stored":260358,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":32418,"internal_metadata":27885918},"hb_peers":[0,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]},{"osd":0,"up_from":12,"seq":51539607616,"num_pgs":1,"num_osds":1,"num_per_pool_osds":1,"num_per_pool_omap_osds":1,"kb":20967424,"kb_used":27940,"kb_used_data":628,"kb_used_omap":34,"kb_used_meta":27229,"kb_avail":20939484,"statfs":{"total":21470642176,"available":21442031616,"internally_reserved":0,"allocated":643072,"data_stored":272016,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":35767,"internal_metadata":27882569},"hb_peers":[1,2,3,4,5,6,7],"snap_trim_queue_len":0,"num_snap_trimming":0,"num_shards_repaired":0,"op_queue_age_hist":{"histogram":[],"upper_bound":1},"perf_stat":{"commit_latency_ms":0,"apply_latency_ms":0,"commit_latency_ns":0,"apply_latency_ns":0},"alerts":[]}],"pool_statfs":[{"poolid":1,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":1,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":462848,"data_stored":459280,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":0,"total":0,"available":0,"internally_reserved":0,"allocated":0,"data_stored":0,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":1,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":2,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":19,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":3,"total":0,"available":0,"internally_reserved":0,"allocated":8192,"data_stored":408,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":5,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0},{"poolid":2,"osd":6,"total":0,"available":0,"internally_reserved":0,"allocated":4096,"data_stored":389,"data_compressed":0,"data_compressed_allocated":0,"data_compressed_original":0,"omap_allocated":0,"internal_metadata":0}]}} 2026-03-21T06:56:04.395 INFO:tasks.cephadm.ceph_manager.ceph:clean! 2026-03-21T06:56:04.395 INFO:tasks.ceph:Waiting until ceph cluster ceph is healthy... 2026-03-21T06:56:04.395 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy 2026-03-21T06:56:04.395 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph health --format=json 2026-03-21T06:56:04.604 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[49158]: debug Successfully removed blocklist entry 2026-03-21T06:56:04.605 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[49158]: debug Removing blocklisted entry for this host : 192.168.123.102:0/1804345957 2026-03-21T06:56:04.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:03.453146+0000 mgr.x (mgr.14152) 348 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-21T06:56:04.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:03.453146+0000 mgr.x (mgr.14152) 348 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-21T06:56:04.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: cephadm 2026-03-21T06:56:03.453907+0000 mgr.x (mgr.14152) 349 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.102:5000 to Dashboard 2026-03-21T06:56:04.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: cephadm 2026-03-21T06:56:03.453907+0000 mgr.x (mgr.14152) 349 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.102:5000 to Dashboard 2026-03-21T06:56:04.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: cephadm 2026-03-21T06:56:03.453973+0000 mgr.x (mgr.14152) 350 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.107:5000 to Dashboard 2026-03-21T06:56:04.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: cephadm 2026-03-21T06:56:03.453973+0000 mgr.x (mgr.14152) 350 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.107:5000 to Dashboard 2026-03-21T06:56:04.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:03.454360+0000 mgr.x (mgr.14152) 351 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-21T06:56:04.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:03.454360+0000 mgr.x (mgr.14152) 351 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-21T06:56:04.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:03.463176+0000 mgr.x (mgr.14152) 352 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm02"}]: dispatch 2026-03-21T06:56:04.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:03.463176+0000 mgr.x (mgr.14152) 352 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm02"}]: dispatch 2026-03-21T06:56:04.605 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:03.469830+0000 mgr.x (mgr.14152) 353 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm07"}]: dispatch 2026-03-21T06:56:04.641 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:04.655 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:03.469830+0000 mgr.x (mgr.14152) 353 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm07"}]: dispatch 2026-03-21T06:56:04.655 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: cluster 2026-03-21T06:56:03.605524+0000 mgr.x (mgr.14152) 354 : cluster [DBG] pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:04.655 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: cluster 2026-03-21T06:56:03.605524+0000 mgr.x (mgr.14152) 354 : cluster [DBG] pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:04.655 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:03.706059+0000 mgr.x (mgr.14152) 355 : audit [DBG] from='client.14664 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:04.655 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:03.706059+0000 mgr.x (mgr.14152) 355 : audit [DBG] from='client.14664 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:04.655 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:04.475995+0000 mon.a (mon.0) 836 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"}]': finished 2026-03-21T06:56:04.655 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: audit 2026-03-21T06:56:04.475995+0000 mon.a (mon.0) 836 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"}]': finished 2026-03-21T06:56:04.655 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: cluster 2026-03-21T06:56:04.478342+0000 mon.a (mon.0) 837 : cluster [DBG] osdmap e67: 8 total, 8 up, 8 in 2026-03-21T06:56:04.655 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:04 vm02 bash[17657]: cluster 2026-03-21T06:56:04.478342+0000 mon.a (mon.0) 837 : cluster [DBG] osdmap e67: 8 total, 8 up, 8 in 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:03.453146+0000 mgr.x (mgr.14152) 348 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:03.453146+0000 mgr.x (mgr.14152) 348 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: cephadm 2026-03-21T06:56:03.453907+0000 mgr.x (mgr.14152) 349 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.102:5000 to Dashboard 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: cephadm 2026-03-21T06:56:03.453907+0000 mgr.x (mgr.14152) 349 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.102:5000 to Dashboard 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: cephadm 2026-03-21T06:56:03.453973+0000 mgr.x (mgr.14152) 350 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.107:5000 to Dashboard 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: cephadm 2026-03-21T06:56:03.453973+0000 mgr.x (mgr.14152) 350 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.107:5000 to Dashboard 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:03.454360+0000 mgr.x (mgr.14152) 351 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:03.454360+0000 mgr.x (mgr.14152) 351 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:03.463176+0000 mgr.x (mgr.14152) 352 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm02"}]: dispatch 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:03.463176+0000 mgr.x (mgr.14152) 352 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm02"}]: dispatch 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:03.469830+0000 mgr.x (mgr.14152) 353 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm07"}]: dispatch 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:03.469830+0000 mgr.x (mgr.14152) 353 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm07"}]: dispatch 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: cluster 2026-03-21T06:56:03.605524+0000 mgr.x (mgr.14152) 354 : cluster [DBG] pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: cluster 2026-03-21T06:56:03.605524+0000 mgr.x (mgr.14152) 354 : cluster [DBG] pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:03.706059+0000 mgr.x (mgr.14152) 355 : audit [DBG] from='client.14664 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:03.706059+0000 mgr.x (mgr.14152) 355 : audit [DBG] from='client.14664 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:04.475995+0000 mon.a (mon.0) 836 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"}]': finished 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: audit 2026-03-21T06:56:04.475995+0000 mon.a (mon.0) 836 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"}]': finished 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: cluster 2026-03-21T06:56:04.478342+0000 mon.a (mon.0) 837 : cluster [DBG] osdmap e67: 8 total, 8 up, 8 in 2026-03-21T06:56:04.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:04 vm07 bash[19945]: cluster 2026-03-21T06:56:04.478342+0000 mon.a (mon.0) 837 : cluster [DBG] osdmap e67: 8 total, 8 up, 8 in 2026-03-21T06:56:04.993 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:56:04.993 INFO:teuthology.orchestra.run.vm02.stdout:{"status":"HEALTH_OK","checks":{},"mutes":[]} 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:03.453146+0000 mgr.x (mgr.14152) 348 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:03.453146+0000 mgr.x (mgr.14152) 348 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-list"}]: dispatch 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: cephadm 2026-03-21T06:56:03.453907+0000 mgr.x (mgr.14152) 349 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.102:5000 to Dashboard 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: cephadm 2026-03-21T06:56:03.453907+0000 mgr.x (mgr.14152) 349 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.102:5000 to Dashboard 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: cephadm 2026-03-21T06:56:03.453973+0000 mgr.x (mgr.14152) 350 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.107:5000 to Dashboard 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: cephadm 2026-03-21T06:56:03.453973+0000 mgr.x (mgr.14152) 350 : cephadm [INF] Adding iSCSI gateway http://:@192.168.123.107:5000 to Dashboard 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:03.454360+0000 mgr.x (mgr.14152) 351 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:03.454360+0000 mgr.x (mgr.14152) 351 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard set-iscsi-api-ssl-verification", "value": "true"}]: dispatch 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:03.463176+0000 mgr.x (mgr.14152) 352 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm02"}]: dispatch 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:03.463176+0000 mgr.x (mgr.14152) 352 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm02"}]: dispatch 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:03.469830+0000 mgr.x (mgr.14152) 353 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm07"}]: dispatch 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:03.469830+0000 mgr.x (mgr.14152) 353 : audit [DBG] from='mon.0 -' entity='mon.' cmd=[{"prefix": "dashboard iscsi-gateway-add", "name": "vm07"}]: dispatch 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: cluster 2026-03-21T06:56:03.605524+0000 mgr.x (mgr.14152) 354 : cluster [DBG] pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: cluster 2026-03-21T06:56:03.605524+0000 mgr.x (mgr.14152) 354 : cluster [DBG] pgmap v293: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:03.706059+0000 mgr.x (mgr.14152) 355 : audit [DBG] from='client.14664 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:03.706059+0000 mgr.x (mgr.14152) 355 : audit [DBG] from='client.14664 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:04.475995+0000 mon.a (mon.0) 836 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"}]': finished 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: audit 2026-03-21T06:56:04.475995+0000 mon.a (mon.0) 836 : audit [INF] from='client.? 192.168.123.102:0/886128506' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/2023406208"}]': finished 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: cluster 2026-03-21T06:56:04.478342+0000 mon.a (mon.0) 837 : cluster [DBG] osdmap e67: 8 total, 8 up, 8 in 2026-03-21T06:56:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:04 vm04 bash[20194]: cluster 2026-03-21T06:56:04.478342+0000 mon.a (mon.0) 837 : cluster [DBG] osdmap e67: 8 total, 8 up, 8 in 2026-03-21T06:56:05.045 INFO:tasks.cephadm.ceph_manager.ceph:wait_until_healthy done 2026-03-21T06:56:05.045 INFO:tasks.cephadm:Setup complete, yielding 2026-03-21T06:56:05.045 INFO:teuthology.run_tasks:Running task cephadm.shell... 2026-03-21T06:56:05.047 INFO:tasks.cephadm:Running commands on role host.a host ubuntu@vm02.local 2026-03-21T06:56:05.047 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- bash -c 'ceph orch status' 2026-03-21T06:56:05.286 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:05.624 INFO:teuthology.orchestra.run.vm02.stdout:Backend: cephadm 2026-03-21T06:56:05.624 INFO:teuthology.orchestra.run.vm02.stdout:Available: Yes 2026-03-21T06:56:05.624 INFO:teuthology.orchestra.run.vm02.stdout:Paused: No 2026-03-21T06:56:05.679 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- bash -c 'ceph orch ps' 2026-03-21T06:56:05.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:05 vm02 bash[17657]: audit 2026-03-21T06:56:04.340656+0000 mgr.x (mgr.14152) 356 : audit [DBG] from='client.14670 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:05.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:05 vm02 bash[17657]: audit 2026-03-21T06:56:04.340656+0000 mgr.x (mgr.14152) 356 : audit [DBG] from='client.14670 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:05.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:05 vm02 bash[17657]: audit 2026-03-21T06:56:04.703644+0000 mon.a (mon.0) 838 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"} : dispatch 2026-03-21T06:56:05.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:05 vm02 bash[17657]: audit 2026-03-21T06:56:04.703644+0000 mon.a (mon.0) 838 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"} : dispatch 2026-03-21T06:56:05.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:05 vm02 bash[17657]: audit 2026-03-21T06:56:04.994707+0000 mon.a (mon.0) 839 : audit [DBG] from='client.? 192.168.123.102:0/2314032958' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-21T06:56:05.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:05 vm02 bash[17657]: audit 2026-03-21T06:56:04.994707+0000 mon.a (mon.0) 839 : audit [DBG] from='client.? 192.168.123.102:0/2314032958' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-21T06:56:05.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:05 vm02 bash[49158]: debug Successfully removed blocklist entry 2026-03-21T06:56:05.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:05 vm02 bash[49158]: debug Removing blocklisted entry for this host : 192.168.123.102:0/288488007 2026-03-21T06:56:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:05 vm07 bash[19945]: audit 2026-03-21T06:56:04.340656+0000 mgr.x (mgr.14152) 356 : audit [DBG] from='client.14670 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:05 vm07 bash[19945]: audit 2026-03-21T06:56:04.340656+0000 mgr.x (mgr.14152) 356 : audit [DBG] from='client.14670 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:05 vm07 bash[19945]: audit 2026-03-21T06:56:04.703644+0000 mon.a (mon.0) 838 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"} : dispatch 2026-03-21T06:56:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:05 vm07 bash[19945]: audit 2026-03-21T06:56:04.703644+0000 mon.a (mon.0) 838 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"} : dispatch 2026-03-21T06:56:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:05 vm07 bash[19945]: audit 2026-03-21T06:56:04.994707+0000 mon.a (mon.0) 839 : audit [DBG] from='client.? 192.168.123.102:0/2314032958' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-21T06:56:05.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:05 vm07 bash[19945]: audit 2026-03-21T06:56:04.994707+0000 mon.a (mon.0) 839 : audit [DBG] from='client.? 192.168.123.102:0/2314032958' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-21T06:56:05.929 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:06.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:05 vm04 bash[20194]: audit 2026-03-21T06:56:04.340656+0000 mgr.x (mgr.14152) 356 : audit [DBG] from='client.14670 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:06.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:05 vm04 bash[20194]: audit 2026-03-21T06:56:04.340656+0000 mgr.x (mgr.14152) 356 : audit [DBG] from='client.14670 -' entity='client.admin' cmd=[{"prefix": "pg dump", "target": ["mon-mgr", ""], "format": "json"}]: dispatch 2026-03-21T06:56:06.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:05 vm04 bash[20194]: audit 2026-03-21T06:56:04.703644+0000 mon.a (mon.0) 838 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"} : dispatch 2026-03-21T06:56:06.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:05 vm04 bash[20194]: audit 2026-03-21T06:56:04.703644+0000 mon.a (mon.0) 838 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"} : dispatch 2026-03-21T06:56:06.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:05 vm04 bash[20194]: audit 2026-03-21T06:56:04.994707+0000 mon.a (mon.0) 839 : audit [DBG] from='client.? 192.168.123.102:0/2314032958' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-21T06:56:06.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:05 vm04 bash[20194]: audit 2026-03-21T06:56:04.994707+0000 mon.a (mon.0) 839 : audit [DBG] from='client.? 192.168.123.102:0/2314032958' entity='client.admin' cmd={"prefix": "health", "format": "json"} : dispatch 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:NAME HOST PORTS STATUS REFRESHED AGE MEM USE MEM LIM VERSION IMAGE ID CONTAINER ID 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:iscsi.iscsi.a vm02 *:5000 running (9s) 3s ago 9s 206M - 3.9 26d640ab1ceb 3366610c7ee8 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:iscsi.iscsi.b vm07 *:5000 running (9s) 3s ago 9s 111M - 3.9 26d640ab1ceb 67ce872d373b 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:mgr.x vm02 *:9283,8765 running (8m) 3s ago 8m 590M - 20.2.0-712-g70f8415b 26d640ab1ceb f0d6b92f6c52 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:mon.a vm02 running (8m) 3s ago 8m 110M 2048M 20.2.0-712-g70f8415b 26d640ab1ceb 3f3ab17ac9a4 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:mon.b vm04 running (7m) 2m ago 7m 103M 2048M 20.2.0-712-g70f8415b 26d640ab1ceb 20b7ad5e1cac 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:mon.c vm07 running (7m) 3s ago 7m 100M 2048M 20.2.0-712-g70f8415b 26d640ab1ceb 351d6d2dc031 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:osd.0 vm02 running (5m) 3s ago 7m 105M 919M 20.2.0-712-g70f8415b 26d640ab1ceb 71c8b9cf073d 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:osd.1 vm02 running (5m) 3s ago 5m 104M 919M 20.2.0-712-g70f8415b 26d640ab1ceb 81795e0c37d0 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:osd.2 vm04 running (5m) 2m ago 5m 103M 2967M 20.2.0-712-g70f8415b 26d640ab1ceb 764da96c1b55 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:osd.3 vm04 running (4m) 2m ago 4m 103M 2967M 20.2.0-712-g70f8415b 26d640ab1ceb 8abdb827a1d9 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:osd.4 vm04 running (4m) 2m ago 4m 100M 2967M 20.2.0-712-g70f8415b 26d640ab1ceb 2c9cbd422c5a 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:osd.5 vm07 running (2m) 3s ago 4m 101M 1978M 20.2.0-712-g70f8415b 26d640ab1ceb ad90132f4dda 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:osd.6 vm07 running (2m) 3s ago 2m 100M 1978M 20.2.0-712-g70f8415b 26d640ab1ceb cc40aba601bd 2026-03-21T06:56:06.260 INFO:teuthology.orchestra.run.vm02.stdout:osd.7 vm07 running (18s) 3s ago 2m 93.9M 1978M 20.2.0-712-g70f8415b 26d640ab1ceb 2a9aec30491a 2026-03-21T06:56:06.317 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- bash -c 'ceph orch ls' 2026-03-21T06:56:06.570 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: debug Successfully removed blocklist entry 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: debug Reading the configuration object to update local LIO configuration 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: debug Configuration does not have an entry for this host(vm02.local) - nothing to define to LIO 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: debug * Running on all addresses. 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: * Running on all addresses. 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: debug * Running on http://[::1]:5000/ (Press CTRL+C to quit) 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: * Running on http://[::1]:5000/ (Press CTRL+C to quit) 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: * Serving Flask app 'rbd-target-api' (lazy loading) 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: * Environment: production 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: WARNING: This is a development server. Do not use it in a production deployment. 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: Use a production WSGI server instead. 2026-03-21T06:56:06.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[49158]: * Debug mode: off 2026-03-21T06:56:06.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[17657]: cluster 2026-03-21T06:56:05.605826+0000 mgr.x (mgr.14152) 357 : cluster [DBG] pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 2 op/s 2026-03-21T06:56:06.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[17657]: cluster 2026-03-21T06:56:05.605826+0000 mgr.x (mgr.14152) 357 : cluster [DBG] pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 2 op/s 2026-03-21T06:56:06.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[17657]: audit 2026-03-21T06:56:05.616199+0000 mon.a (mon.0) 840 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"}]': finished 2026-03-21T06:56:06.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[17657]: audit 2026-03-21T06:56:05.616199+0000 mon.a (mon.0) 840 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"}]': finished 2026-03-21T06:56:06.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[17657]: cluster 2026-03-21T06:56:05.617907+0000 mon.a (mon.0) 841 : cluster [DBG] osdmap e68: 8 total, 8 up, 8 in 2026-03-21T06:56:06.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[17657]: cluster 2026-03-21T06:56:05.617907+0000 mon.a (mon.0) 841 : cluster [DBG] osdmap e68: 8 total, 8 up, 8 in 2026-03-21T06:56:06.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[17657]: audit 2026-03-21T06:56:05.623600+0000 mgr.x (mgr.14152) 358 : audit [DBG] from='client.24589 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:06.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[17657]: audit 2026-03-21T06:56:05.623600+0000 mgr.x (mgr.14152) 358 : audit [DBG] from='client.24589 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:06.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[17657]: audit 2026-03-21T06:56:05.848804+0000 mon.a (mon.0) 842 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"} : dispatch 2026-03-21T06:56:06.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:06 vm02 bash[17657]: audit 2026-03-21T06:56:05.848804+0000 mon.a (mon.0) 842 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"} : dispatch 2026-03-21T06:56:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:06 vm07 bash[19945]: cluster 2026-03-21T06:56:05.605826+0000 mgr.x (mgr.14152) 357 : cluster [DBG] pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 2 op/s 2026-03-21T06:56:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:06 vm07 bash[19945]: cluster 2026-03-21T06:56:05.605826+0000 mgr.x (mgr.14152) 357 : cluster [DBG] pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 2 op/s 2026-03-21T06:56:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:06 vm07 bash[19945]: audit 2026-03-21T06:56:05.616199+0000 mon.a (mon.0) 840 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"}]': finished 2026-03-21T06:56:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:06 vm07 bash[19945]: audit 2026-03-21T06:56:05.616199+0000 mon.a (mon.0) 840 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"}]': finished 2026-03-21T06:56:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:06 vm07 bash[19945]: cluster 2026-03-21T06:56:05.617907+0000 mon.a (mon.0) 841 : cluster [DBG] osdmap e68: 8 total, 8 up, 8 in 2026-03-21T06:56:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:06 vm07 bash[19945]: cluster 2026-03-21T06:56:05.617907+0000 mon.a (mon.0) 841 : cluster [DBG] osdmap e68: 8 total, 8 up, 8 in 2026-03-21T06:56:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:06 vm07 bash[19945]: audit 2026-03-21T06:56:05.623600+0000 mgr.x (mgr.14152) 358 : audit [DBG] from='client.24589 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:06 vm07 bash[19945]: audit 2026-03-21T06:56:05.623600+0000 mgr.x (mgr.14152) 358 : audit [DBG] from='client.24589 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:06 vm07 bash[19945]: audit 2026-03-21T06:56:05.848804+0000 mon.a (mon.0) 842 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"} : dispatch 2026-03-21T06:56:06.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:06 vm07 bash[19945]: audit 2026-03-21T06:56:05.848804+0000 mon.a (mon.0) 842 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"} : dispatch 2026-03-21T06:56:06.929 INFO:teuthology.orchestra.run.vm02.stdout:NAME PORTS RUNNING REFRESHED AGE PLACEMENT 2026-03-21T06:56:06.929 INFO:teuthology.orchestra.run.vm02.stdout:iscsi.datapool ?:5000 2/2 4s ago 13s vm02=iscsi.a;vm07=iscsi.b;count:2 2026-03-21T06:56:06.929 INFO:teuthology.orchestra.run.vm02.stdout:mgr 1/1 4s ago 7m vm02=x;count:1 2026-03-21T06:56:06.929 INFO:teuthology.orchestra.run.vm02.stdout:mon 3/3 2m ago 7m vm02:192.168.123.102=a;vm04:192.168.123.104=b;vm07:192.168.123.107=c;count:3 2026-03-21T06:56:06.929 INFO:teuthology.orchestra.run.vm02.stdout:osd.default 8 2m ago 2m vm07 2026-03-21T06:56:06.987 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- bash -c 'ceph orch host ls' 2026-03-21T06:56:07.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:06 vm04 bash[20194]: cluster 2026-03-21T06:56:05.605826+0000 mgr.x (mgr.14152) 357 : cluster [DBG] pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 2 op/s 2026-03-21T06:56:07.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:06 vm04 bash[20194]: cluster 2026-03-21T06:56:05.605826+0000 mgr.x (mgr.14152) 357 : cluster [DBG] pgmap v295: 4 pgs: 4 active+clean; 449 KiB data, 218 MiB used, 160 GiB / 160 GiB avail; 2.2 KiB/s rd, 2 op/s 2026-03-21T06:56:07.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:06 vm04 bash[20194]: audit 2026-03-21T06:56:05.616199+0000 mon.a (mon.0) 840 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"}]': finished 2026-03-21T06:56:07.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:06 vm04 bash[20194]: audit 2026-03-21T06:56:05.616199+0000 mon.a (mon.0) 840 : audit [INF] from='client.? 192.168.123.102:0/3423810180' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/1804345957"}]': finished 2026-03-21T06:56:07.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:06 vm04 bash[20194]: cluster 2026-03-21T06:56:05.617907+0000 mon.a (mon.0) 841 : cluster [DBG] osdmap e68: 8 total, 8 up, 8 in 2026-03-21T06:56:07.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:06 vm04 bash[20194]: cluster 2026-03-21T06:56:05.617907+0000 mon.a (mon.0) 841 : cluster [DBG] osdmap e68: 8 total, 8 up, 8 in 2026-03-21T06:56:07.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:06 vm04 bash[20194]: audit 2026-03-21T06:56:05.623600+0000 mgr.x (mgr.14152) 358 : audit [DBG] from='client.24589 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:07.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:06 vm04 bash[20194]: audit 2026-03-21T06:56:05.623600+0000 mgr.x (mgr.14152) 358 : audit [DBG] from='client.24589 -' entity='client.admin' cmd=[{"prefix": "orch status", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:07.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:06 vm04 bash[20194]: audit 2026-03-21T06:56:05.848804+0000 mon.a (mon.0) 842 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"} : dispatch 2026-03-21T06:56:07.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:06 vm04 bash[20194]: audit 2026-03-21T06:56:05.848804+0000 mon.a (mon.0) 842 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"} : dispatch 2026-03-21T06:56:07.234 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:07.552 INFO:teuthology.orchestra.run.vm02.stdout:HOST ADDR LABELS STATUS 2026-03-21T06:56:07.552 INFO:teuthology.orchestra.run.vm02.stdout:vm02 192.168.123.102 2026-03-21T06:56:07.552 INFO:teuthology.orchestra.run.vm02.stdout:vm04 192.168.123.104 2026-03-21T06:56:07.552 INFO:teuthology.orchestra.run.vm02.stdout:vm07 192.168.123.107 2026-03-21T06:56:07.553 INFO:teuthology.orchestra.run.vm02.stdout:3 hosts in cluster 2026-03-21T06:56:07.639 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- bash -c 'ceph orch device ls' 2026-03-21T06:56:07.880 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T06:56:07.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:07 vm02 bash[17657]: audit 2026-03-21T06:56:06.258488+0000 mgr.x (mgr.14152) 359 : audit [DBG] from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:07.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:07 vm02 bash[17657]: audit 2026-03-21T06:56:06.258488+0000 mgr.x (mgr.14152) 359 : audit [DBG] from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:07.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:07 vm02 bash[17657]: audit 2026-03-21T06:56:06.623354+0000 mgr.x (mgr.14152) 360 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:07.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:07 vm02 bash[17657]: audit 2026-03-21T06:56:06.623354+0000 mgr.x (mgr.14152) 360 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:07.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:07 vm02 bash[17657]: audit 2026-03-21T06:56:06.662393+0000 mon.a (mon.0) 843 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"}]': finished 2026-03-21T06:56:07.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:07 vm02 bash[17657]: audit 2026-03-21T06:56:06.662393+0000 mon.a (mon.0) 843 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"}]': finished 2026-03-21T06:56:07.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:07 vm02 bash[17657]: cluster 2026-03-21T06:56:06.664929+0000 mon.a (mon.0) 844 : cluster [DBG] osdmap e69: 8 total, 8 up, 8 in 2026-03-21T06:56:07.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:07 vm02 bash[17657]: cluster 2026-03-21T06:56:06.664929+0000 mon.a (mon.0) 844 : cluster [DBG] osdmap e69: 8 total, 8 up, 8 in 2026-03-21T06:56:07.901 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:07 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:56:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:07 vm07 bash[19945]: audit 2026-03-21T06:56:06.258488+0000 mgr.x (mgr.14152) 359 : audit [DBG] from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:07 vm07 bash[19945]: audit 2026-03-21T06:56:06.258488+0000 mgr.x (mgr.14152) 359 : audit [DBG] from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:07 vm07 bash[19945]: audit 2026-03-21T06:56:06.623354+0000 mgr.x (mgr.14152) 360 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:07 vm07 bash[19945]: audit 2026-03-21T06:56:06.623354+0000 mgr.x (mgr.14152) 360 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:07 vm07 bash[19945]: audit 2026-03-21T06:56:06.662393+0000 mon.a (mon.0) 843 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"}]': finished 2026-03-21T06:56:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:07 vm07 bash[19945]: audit 2026-03-21T06:56:06.662393+0000 mon.a (mon.0) 843 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"}]': finished 2026-03-21T06:56:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:07 vm07 bash[19945]: cluster 2026-03-21T06:56:06.664929+0000 mon.a (mon.0) 844 : cluster [DBG] osdmap e69: 8 total, 8 up, 8 in 2026-03-21T06:56:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:07 vm07 bash[19945]: cluster 2026-03-21T06:56:06.664929+0000 mon.a (mon.0) 844 : cluster [DBG] osdmap e69: 8 total, 8 up, 8 in 2026-03-21T06:56:08.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:07 vm04 bash[20194]: audit 2026-03-21T06:56:06.258488+0000 mgr.x (mgr.14152) 359 : audit [DBG] from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:08.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:07 vm04 bash[20194]: audit 2026-03-21T06:56:06.258488+0000 mgr.x (mgr.14152) 359 : audit [DBG] from='client.14700 -' entity='client.admin' cmd=[{"prefix": "orch ps", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:08.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:07 vm04 bash[20194]: audit 2026-03-21T06:56:06.623354+0000 mgr.x (mgr.14152) 360 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:08.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:07 vm04 bash[20194]: audit 2026-03-21T06:56:06.623354+0000 mgr.x (mgr.14152) 360 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:08.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:07 vm04 bash[20194]: audit 2026-03-21T06:56:06.662393+0000 mon.a (mon.0) 843 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"}]': finished 2026-03-21T06:56:08.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:07 vm04 bash[20194]: audit 2026-03-21T06:56:06.662393+0000 mon.a (mon.0) 843 : audit [INF] from='client.? 192.168.123.102:0/2370027900' entity='client.iscsi.iscsi.a' cmd='[{"prefix": "osd blocklist", "blocklistop": "rm", "addr": "192.168.123.102:0/288488007"}]': finished 2026-03-21T06:56:08.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:07 vm04 bash[20194]: cluster 2026-03-21T06:56:06.664929+0000 mon.a (mon.0) 844 : cluster [DBG] osdmap e69: 8 total, 8 up, 8 in 2026-03-21T06:56:08.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:07 vm04 bash[20194]: cluster 2026-03-21T06:56:06.664929+0000 mon.a (mon.0) 844 : cluster [DBG] osdmap e69: 8 total, 8 up, 8 in 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:HOST PATH TYPE DEVICE ID SIZE AVAILABLE REFRESHED REJECT REASONS 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm02 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 2m ago Has a FileSystem, Insufficient space (<5GB) 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm02 /dev/vdb hdd DWNBRSTVMM02001 20.0G No 2m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm02 /dev/vdc hdd DWNBRSTVMM02002 20.0G No 2m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm02 /dev/vdd hdd DWNBRSTVMM02003 20.0G No 2m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm02 /dev/vde hdd DWNBRSTVMM02004 20.0G No 2m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm04 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 2m ago Has a FileSystem, Insufficient space (<5GB) 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm04 /dev/vdb hdd DWNBRSTVMM04001 20.0G No 2m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm04 /dev/vdc hdd DWNBRSTVMM04002 20.0G No 2m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm04 /dev/vdd hdd DWNBRSTVMM04003 20.0G No 2m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm04 /dev/vde hdd DWNBRSTVMM04004 20.0G No 2m ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm07 /dev/sr0 hdd QEMU_DVD-ROM_QM00003 366k No 4s ago Has a FileSystem, Insufficient space (<5GB) 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm07 /dev/vdb hdd DWNBRSTVMM07001 20.0G No 4s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm07 /dev/vdc hdd DWNBRSTVMM07002 20.0G No 4s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm07 /dev/vdd hdd DWNBRSTVMM07003 20.0G No 4s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.216 INFO:teuthology.orchestra.run.vm02.stdout:vm07 /dev/vde hdd DWNBRSTVMM07004 20.0G No 4s ago Has a FileSystem, Insufficient space (<10 extents) on vgs, LVM detected 2026-03-21T06:56:08.269 INFO:teuthology.run_tasks:Running task install... 2026-03-21T06:56:08.271 DEBUG:teuthology.task.install:project ceph 2026-03-21T06:56:08.271 DEBUG:teuthology.task.install:INSTALL overrides: {'ceph': {'flavor': 'default', 'sha1': '70f8415b300f041766fa27faf7d5472699e32388'}, 'extra_system_packages': {'deb': ['python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}} 2026-03-21T06:56:08.271 DEBUG:teuthology.task.install:config {'extra_system_packages': {'deb': ['open-iscsi', 'multipath-tools', 'python3-jmespath', 'python3-xmltodict', 's3cmd'], 'rpm': ['iscsi-initiator-utils', 'device-mapper-multipath', 'bzip2', 'perl-Test-Harness', 'python3-jmespath', 'python3-xmltodict', 's3cmd']}, 'flavor': 'default', 'sha1': '70f8415b300f041766fa27faf7d5472699e32388'} 2026-03-21T06:56:08.271 INFO:teuthology.task.install:Using flavor: default 2026-03-21T06:56:08.273 DEBUG:teuthology.task.install:Package list is: {'deb': ['ceph', 'cephadm', 'ceph-mds', 'ceph-mgr', 'ceph-common', 'ceph-fuse', 'ceph-test', 'ceph-volume', 'radosgw', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'libcephfs2', 'libcephfs-dev', 'librados2', 'librbd1', 'rbd-fuse'], 'rpm': ['ceph-radosgw', 'ceph-test', 'ceph', 'ceph-base', 'cephadm', 'ceph-immutable-object-cache', 'ceph-mgr', 'ceph-mgr-dashboard', 'ceph-mgr-diskprediction-local', 'ceph-mgr-rook', 'ceph-mgr-cephadm', 'ceph-fuse', 'ceph-volume', 'librados-devel', 'libcephfs2', 'libcephfs-devel', 'librados2', 'librbd1', 'python3-rados', 'python3-rgw', 'python3-cephfs', 'python3-rbd', 'rbd-fuse', 'rbd-mirror', 'rbd-nbd']} 2026-03-21T06:56:08.273 INFO:teuthology.task.install:extra packages: [] 2026-03-21T06:56:08.273 DEBUG:teuthology.orchestra.run.vm02:> sudo apt-key list | grep Ceph 2026-03-21T06:56:08.273 DEBUG:teuthology.orchestra.run.vm04:> sudo apt-key list | grep Ceph 2026-03-21T06:56:08.274 DEBUG:teuthology.orchestra.run.vm07:> sudo apt-key list | grep Ceph 2026-03-21T06:56:08.313 INFO:teuthology.orchestra.run.vm02.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-21T06:56:08.315 INFO:teuthology.orchestra.run.vm07.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-21T06:56:08.316 INFO:teuthology.orchestra.run.vm04.stderr:Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). 2026-03-21T06:56:08.336 INFO:teuthology.orchestra.run.vm02.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-21T06:56:08.336 INFO:teuthology.orchestra.run.vm02.stdout:uid [ unknown] Ceph.com (release key) 2026-03-21T06:56:08.336 INFO:teuthology.orchestra.run.vm07.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-21T06:56:08.336 INFO:teuthology.orchestra.run.vm07.stdout:uid [ unknown] Ceph.com (release key) 2026-03-21T06:56:08.336 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-21T06:56:08.336 INFO:teuthology.task.install.deb:Installing system (non-project) packages: open-iscsi, multipath-tools, python3-jmespath, python3-xmltodict, s3cmd on remote deb x86_64 2026-03-21T06:56:08.336 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-21T06:56:08.372 INFO:teuthology.orchestra.run.vm04.stdout:uid [ unknown] Ceph automated package build (Ceph automated package build) 2026-03-21T06:56:08.372 INFO:teuthology.orchestra.run.vm04.stdout:uid [ unknown] Ceph.com (release key) 2026-03-21T06:56:08.372 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-21T06:56:08.372 INFO:teuthology.task.install.deb:Installing system (non-project) packages: open-iscsi, multipath-tools, python3-jmespath, python3-xmltodict, s3cmd on remote deb x86_64 2026-03-21T06:56:08.372 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-21T06:56:08.372 INFO:teuthology.task.install.deb:Installing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on remote deb x86_64 2026-03-21T06:56:08.372 INFO:teuthology.task.install.deb:Installing system (non-project) packages: open-iscsi, multipath-tools, python3-jmespath, python3-xmltodict, s3cmd on remote deb x86_64 2026-03-21T06:56:08.373 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-21T06:56:08.944 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default/ 2026-03-21T06:56:08.944 INFO:teuthology.task.install.deb:Package version is 20.2.0-712-g70f8415b-1jammy 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: audit 2026-03-21T06:56:06.929016+0000 mgr.x (mgr.14152) 361 : audit [DBG] from='client.14706 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: audit 2026-03-21T06:56:06.929016+0000 mgr.x (mgr.14152) 361 : audit [DBG] from='client.14706 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: audit 2026-03-21T06:56:07.459792+0000 mgr.x (mgr.14152) 362 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: audit 2026-03-21T06:56:07.459792+0000 mgr.x (mgr.14152) 362 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: audit 2026-03-21T06:56:07.553817+0000 mgr.x (mgr.14152) 363 : audit [DBG] from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: audit 2026-03-21T06:56:07.553817+0000 mgr.x (mgr.14152) 363 : audit [DBG] from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: cluster 2026-03-21T06:56:07.606169+0000 mgr.x (mgr.14152) 364 : cluster [DBG] pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 2 op/s 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: cluster 2026-03-21T06:56:07.606169+0000 mgr.x (mgr.14152) 364 : cluster [DBG] pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 2 op/s 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: audit 2026-03-21T06:56:08.506935+0000 mon.a (mon.0) 845 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: audit 2026-03-21T06:56:08.506935+0000 mon.a (mon.0) 845 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: audit 2026-03-21T06:56:08.511778+0000 mon.a (mon.0) 846 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:08 vm04 bash[20194]: audit 2026-03-21T06:56:08.511778+0000 mon.a (mon.0) 846 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.044 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default/ 2026-03-21T06:56:09.044 INFO:teuthology.task.install.deb:Package version is 20.2.0-712-g70f8415b-1jammy 2026-03-21T06:56:09.131 INFO:teuthology.task.install.deb:Pulling from https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default/ 2026-03-21T06:56:09.132 INFO:teuthology.task.install.deb:Package version is 20.2.0-712-g70f8415b-1jammy 2026-03-21T06:56:09.169 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: audit 2026-03-21T06:56:06.929016+0000 mgr.x (mgr.14152) 361 : audit [DBG] from='client.14706 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.169 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: audit 2026-03-21T06:56:06.929016+0000 mgr.x (mgr.14152) 361 : audit [DBG] from='client.14706 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.169 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: audit 2026-03-21T06:56:07.459792+0000 mgr.x (mgr.14152) 362 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:09.169 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: audit 2026-03-21T06:56:07.459792+0000 mgr.x (mgr.14152) 362 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:09.169 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: audit 2026-03-21T06:56:07.553817+0000 mgr.x (mgr.14152) 363 : audit [DBG] from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.169 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: audit 2026-03-21T06:56:07.553817+0000 mgr.x (mgr.14152) 363 : audit [DBG] from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.169 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: cluster 2026-03-21T06:56:07.606169+0000 mgr.x (mgr.14152) 364 : cluster [DBG] pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 2 op/s 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: cluster 2026-03-21T06:56:07.606169+0000 mgr.x (mgr.14152) 364 : cluster [DBG] pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 2 op/s 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: audit 2026-03-21T06:56:08.506935+0000 mon.a (mon.0) 845 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: audit 2026-03-21T06:56:08.506935+0000 mon.a (mon.0) 845 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: audit 2026-03-21T06:56:08.511778+0000 mon.a (mon.0) 846 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:08 vm07 bash[19945]: audit 2026-03-21T06:56:08.511778+0000 mon.a (mon.0) 846 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: audit 2026-03-21T06:56:06.929016+0000 mgr.x (mgr.14152) 361 : audit [DBG] from='client.14706 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: audit 2026-03-21T06:56:06.929016+0000 mgr.x (mgr.14152) 361 : audit [DBG] from='client.14706 -' entity='client.admin' cmd=[{"prefix": "orch ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: audit 2026-03-21T06:56:07.459792+0000 mgr.x (mgr.14152) 362 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: audit 2026-03-21T06:56:07.459792+0000 mgr.x (mgr.14152) 362 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: audit 2026-03-21T06:56:07.553817+0000 mgr.x (mgr.14152) 363 : audit [DBG] from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: audit 2026-03-21T06:56:07.553817+0000 mgr.x (mgr.14152) 363 : audit [DBG] from='client.14712 -' entity='client.admin' cmd=[{"prefix": "orch host ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: cluster 2026-03-21T06:56:07.606169+0000 mgr.x (mgr.14152) 364 : cluster [DBG] pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 2 op/s 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: cluster 2026-03-21T06:56:07.606169+0000 mgr.x (mgr.14152) 364 : cluster [DBG] pgmap v298: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.3 KiB/s rd, 2 op/s 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: audit 2026-03-21T06:56:08.506935+0000 mon.a (mon.0) 845 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: audit 2026-03-21T06:56:08.506935+0000 mon.a (mon.0) 845 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: audit 2026-03-21T06:56:08.511778+0000 mon.a (mon.0) 846 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.170 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:08 vm02 bash[17657]: audit 2026-03-21T06:56:08.511778+0000 mon.a (mon.0) 846 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.468 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:56:09.468 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-21T06:56:09.476 DEBUG:teuthology.orchestra.run.vm02:> sudo apt-get update 2026-03-21T06:56:09.581 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:56:09.581 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-21T06:56:09.589 DEBUG:teuthology.orchestra.run.vm07:> sudo apt-get update 2026-03-21T06:56:09.628 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:56:09.628 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/apt/sources.list.d/ceph.list 2026-03-21T06:56:09.637 DEBUG:teuthology.orchestra.run.vm04:> sudo apt-get update 2026-03-21T06:56:09.650 INFO:teuthology.orchestra.run.vm02.stdout:Hit:1 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-21T06:56:09.749 INFO:teuthology.orchestra.run.vm04.stdout:Hit:1 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:08.216068+0000 mgr.x (mgr.14152) 365 : audit [DBG] from='client.14718 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:08.216068+0000 mgr.x (mgr.14152) 365 : audit [DBG] from='client.14718 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.553526+0000 mon.a (mon.0) 847 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.553526+0000 mon.a (mon.0) 847 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.565542+0000 mon.a (mon.0) 848 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.565542+0000 mon.a (mon.0) 848 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.569869+0000 mon.a (mon.0) 849 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.569869+0000 mon.a (mon.0) 849 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.570821+0000 mon.a (mon.0) 850 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.570821+0000 mon.a (mon.0) 850 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.579937+0000 mon.a (mon.0) 851 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.579937+0000 mon.a (mon.0) 851 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.752 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.581010+0000 mon.a (mon.0) 852 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:09.753 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:09 vm07 bash[19945]: audit 2026-03-21T06:56:09.581010+0000 mon.a (mon.0) 852 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:09.753 INFO:teuthology.orchestra.run.vm04.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-21T06:56:09.761 INFO:teuthology.orchestra.run.vm04.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-21T06:56:09.763 INFO:teuthology.orchestra.run.vm07.stdout:Hit:1 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-21T06:56:09.803 INFO:teuthology.orchestra.run.vm04.stdout:Hit:4 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-21T06:56:09.803 INFO:teuthology.orchestra.run.vm02.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-21T06:56:09.885 INFO:teuthology.orchestra.run.vm07.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:08.216068+0000 mgr.x (mgr.14152) 365 : audit [DBG] from='client.14718 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:08.216068+0000 mgr.x (mgr.14152) 365 : audit [DBG] from='client.14718 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.553526+0000 mon.a (mon.0) 847 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.553526+0000 mon.a (mon.0) 847 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.565542+0000 mon.a (mon.0) 848 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.565542+0000 mon.a (mon.0) 848 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.569869+0000 mon.a (mon.0) 849 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.569869+0000 mon.a (mon.0) 849 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.570821+0000 mon.a (mon.0) 850 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.570821+0000 mon.a (mon.0) 850 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.579937+0000 mon.a (mon.0) 851 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.579937+0000 mon.a (mon.0) 851 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.581010+0000 mon.a (mon.0) 852 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:09.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:09 vm02 bash[17657]: audit 2026-03-21T06:56:09.581010+0000 mon.a (mon.0) 852 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:09.918 INFO:teuthology.orchestra.run.vm02.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-21T06:56:09.987 INFO:teuthology.orchestra.run.vm07.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:08.216068+0000 mgr.x (mgr.14152) 365 : audit [DBG] from='client.14718 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:08.216068+0000 mgr.x (mgr.14152) 365 : audit [DBG] from='client.14718 -' entity='client.admin' cmd=[{"prefix": "orch device ls", "target": ["mon-mgr", ""]}]: dispatch 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.553526+0000 mon.a (mon.0) 847 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.553526+0000 mon.a (mon.0) 847 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.565542+0000 mon.a (mon.0) 848 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.565542+0000 mon.a (mon.0) 848 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.569869+0000 mon.a (mon.0) 849 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.569869+0000 mon.a (mon.0) 849 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.570821+0000 mon.a (mon.0) 850 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.570821+0000 mon.a (mon.0) 850 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.579937+0000 mon.a (mon.0) 851 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.579937+0000 mon.a (mon.0) 851 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.581010+0000 mon.a (mon.0) 852 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:10.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:09 vm04 bash[20194]: audit 2026-03-21T06:56:09.581010+0000 mon.a (mon.0) 852 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:56:10.033 INFO:teuthology.orchestra.run.vm02.stdout:Hit:4 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-21T06:56:10.033 INFO:teuthology.orchestra.run.vm02.stdout:Ign:5 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy InRelease 2026-03-21T06:56:10.089 INFO:teuthology.orchestra.run.vm07.stdout:Hit:4 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-21T06:56:10.146 INFO:teuthology.orchestra.run.vm02.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release [7680 B] 2026-03-21T06:56:10.160 INFO:teuthology.orchestra.run.vm07.stdout:Ign:5 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy InRelease 2026-03-21T06:56:10.186 INFO:teuthology.orchestra.run.vm04.stdout:Ign:5 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy InRelease 2026-03-21T06:56:10.260 INFO:teuthology.orchestra.run.vm02.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-21T06:56:10.277 INFO:teuthology.orchestra.run.vm07.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release [7680 B] 2026-03-21T06:56:10.299 INFO:teuthology.orchestra.run.vm04.stdout:Get:6 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release [7680 B] 2026-03-21T06:56:10.374 INFO:teuthology.orchestra.run.vm02.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.8 kB] 2026-03-21T06:56:10.395 INFO:teuthology.orchestra.run.vm07.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-21T06:56:10.412 INFO:teuthology.orchestra.run.vm04.stdout:Ign:7 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy Release.gpg 2026-03-21T06:56:10.513 INFO:teuthology.orchestra.run.vm07.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.8 kB] 2026-03-21T06:56:10.525 INFO:teuthology.orchestra.run.vm04.stdout:Get:8 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 Packages [18.8 kB] 2026-03-21T06:56:10.560 INFO:teuthology.orchestra.run.vm02.stdout:Fetched 26.5 kB in 1s (28.5 kB/s) 2026-03-21T06:56:10.593 INFO:teuthology.orchestra.run.vm07.stdout:Fetched 26.5 kB in 1s (31.4 kB/s) 2026-03-21T06:56:10.718 INFO:teuthology.orchestra.run.vm04.stdout:Fetched 26.5 kB in 1s (28.7 kB/s) 2026-03-21T06:56:11.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:10 vm04 bash[20194]: cluster 2026-03-21T06:56:09.606486+0000 mgr.x (mgr.14152) 366 : cluster [DBG] pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:11.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:10 vm04 bash[20194]: cluster 2026-03-21T06:56:09.606486+0000 mgr.x (mgr.14152) 366 : cluster [DBG] pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:11.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:10 vm02 bash[17657]: cluster 2026-03-21T06:56:09.606486+0000 mgr.x (mgr.14152) 366 : cluster [DBG] pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:11.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:10 vm02 bash[17657]: cluster 2026-03-21T06:56:09.606486+0000 mgr.x (mgr.14152) 366 : cluster [DBG] pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:10 vm07 bash[19945]: cluster 2026-03-21T06:56:09.606486+0000 mgr.x (mgr.14152) 366 : cluster [DBG] pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:10 vm07 bash[19945]: cluster 2026-03-21T06:56:09.606486+0000 mgr.x (mgr.14152) 366 : cluster [DBG] pgmap v299: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:11.212 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T06:56:11.223 DEBUG:teuthology.orchestra.run.vm02:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=20.2.0-712-g70f8415b-1jammy cephadm=20.2.0-712-g70f8415b-1jammy ceph-mds=20.2.0-712-g70f8415b-1jammy ceph-mgr=20.2.0-712-g70f8415b-1jammy ceph-common=20.2.0-712-g70f8415b-1jammy ceph-fuse=20.2.0-712-g70f8415b-1jammy ceph-test=20.2.0-712-g70f8415b-1jammy ceph-volume=20.2.0-712-g70f8415b-1jammy radosgw=20.2.0-712-g70f8415b-1jammy python3-rados=20.2.0-712-g70f8415b-1jammy python3-rgw=20.2.0-712-g70f8415b-1jammy python3-cephfs=20.2.0-712-g70f8415b-1jammy python3-rbd=20.2.0-712-g70f8415b-1jammy libcephfs2=20.2.0-712-g70f8415b-1jammy libcephfs-dev=20.2.0-712-g70f8415b-1jammy librados2=20.2.0-712-g70f8415b-1jammy librbd1=20.2.0-712-g70f8415b-1jammy rbd-fuse=20.2.0-712-g70f8415b-1jammy 2026-03-21T06:56:11.257 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T06:56:11.262 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T06:56:11.275 DEBUG:teuthology.orchestra.run.vm07:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=20.2.0-712-g70f8415b-1jammy cephadm=20.2.0-712-g70f8415b-1jammy ceph-mds=20.2.0-712-g70f8415b-1jammy ceph-mgr=20.2.0-712-g70f8415b-1jammy ceph-common=20.2.0-712-g70f8415b-1jammy ceph-fuse=20.2.0-712-g70f8415b-1jammy ceph-test=20.2.0-712-g70f8415b-1jammy ceph-volume=20.2.0-712-g70f8415b-1jammy radosgw=20.2.0-712-g70f8415b-1jammy python3-rados=20.2.0-712-g70f8415b-1jammy python3-rgw=20.2.0-712-g70f8415b-1jammy python3-cephfs=20.2.0-712-g70f8415b-1jammy python3-rbd=20.2.0-712-g70f8415b-1jammy libcephfs2=20.2.0-712-g70f8415b-1jammy libcephfs-dev=20.2.0-712-g70f8415b-1jammy librados2=20.2.0-712-g70f8415b-1jammy librbd1=20.2.0-712-g70f8415b-1jammy rbd-fuse=20.2.0-712-g70f8415b-1jammy 2026-03-21T06:56:11.310 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T06:56:11.385 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T06:56:11.399 DEBUG:teuthology.orchestra.run.vm04:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install ceph=20.2.0-712-g70f8415b-1jammy cephadm=20.2.0-712-g70f8415b-1jammy ceph-mds=20.2.0-712-g70f8415b-1jammy ceph-mgr=20.2.0-712-g70f8415b-1jammy ceph-common=20.2.0-712-g70f8415b-1jammy ceph-fuse=20.2.0-712-g70f8415b-1jammy ceph-test=20.2.0-712-g70f8415b-1jammy ceph-volume=20.2.0-712-g70f8415b-1jammy radosgw=20.2.0-712-g70f8415b-1jammy python3-rados=20.2.0-712-g70f8415b-1jammy python3-rgw=20.2.0-712-g70f8415b-1jammy python3-cephfs=20.2.0-712-g70f8415b-1jammy python3-rbd=20.2.0-712-g70f8415b-1jammy libcephfs2=20.2.0-712-g70f8415b-1jammy libcephfs-dev=20.2.0-712-g70f8415b-1jammy librados2=20.2.0-712-g70f8415b-1jammy librbd1=20.2.0-712-g70f8415b-1jammy rbd-fuse=20.2.0-712-g70f8415b-1jammy 2026-03-21T06:56:11.434 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T06:56:11.439 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T06:56:11.440 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T06:56:11.481 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T06:56:11.482 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T06:56:11.583 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T06:56:11.584 INFO:teuthology.orchestra.run.vm02.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T06:56:11.584 INFO:teuthology.orchestra.run.vm02.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-21T06:56:11.584 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout:The following additional packages will be installed: 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-daemon libcephfs-proxy2 libdouble-conversion3 libfuse2 libjq1 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: liblttng-ust1 libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 librgw2 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph libthrift-0.16.0 nvme-cli python-asyncssh-doc 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: python3-iniconfig python3-jaraco.classes python3-jaraco.collections 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: python3-kubernetes python3-natsort python3-pluggy python3-portend 2026-03-21T06:56:11.585 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout: python3-pytest python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout: python3-tempora python3-threadpoolctl python3-toml python3-wcwidth 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout: python3-webob python3-websocket python3-zc.lockfile qttranslations5-l10n 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout:Suggested packages: 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout: python3-influxdb liblua5.3-dev luarocks python-natsort-doc python-psutil-doc 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout: subversion python-pygments-doc ttf-bitstream-vera python3-paste python3-dap 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout: python-sklearn-doc ipython3 python-webob-doc gsmartcontrol smart-notifier 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout: mailx | mailutils 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout:Recommended packages: 2026-03-21T06:56:11.586 INFO:teuthology.orchestra.run.vm02.stdout: btrfs-tools 2026-03-21T06:56:11.618 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T06:56:11.618 INFO:teuthology.orchestra.run.vm07.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T06:56:11.618 INFO:teuthology.orchestra.run.vm07.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-21T06:56:11.618 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T06:56:11.618 INFO:teuthology.orchestra.run.vm07.stdout:The following additional packages will be installed: 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: libcephfs-daemon libcephfs-proxy2 libdouble-conversion3 libfuse2 libjq1 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: liblttng-ust1 libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 librgw2 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: libsqlite3-mod-ceph libthrift-0.16.0 nvme-cli python-asyncssh-doc 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: python3-iniconfig python3-jaraco.classes python3-jaraco.collections 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: python3-kubernetes python3-natsort python3-pluggy python3-portend 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: python3-pytest python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: python3-tempora python3-threadpoolctl python3-toml python3-wcwidth 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: python3-webob python3-websocket python3-zc.lockfile qttranslations5-l10n 2026-03-21T06:56:11.619 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T06:56:11.620 INFO:teuthology.orchestra.run.vm07.stdout:Suggested packages: 2026-03-21T06:56:11.620 INFO:teuthology.orchestra.run.vm07.stdout: python3-influxdb liblua5.3-dev luarocks python-natsort-doc python-psutil-doc 2026-03-21T06:56:11.620 INFO:teuthology.orchestra.run.vm07.stdout: subversion python-pygments-doc ttf-bitstream-vera python3-paste python3-dap 2026-03-21T06:56:11.620 INFO:teuthology.orchestra.run.vm07.stdout: python-sklearn-doc ipython3 python-webob-doc gsmartcontrol smart-notifier 2026-03-21T06:56:11.620 INFO:teuthology.orchestra.run.vm07.stdout: mailx | mailutils 2026-03-21T06:56:11.620 INFO:teuthology.orchestra.run.vm07.stdout:Recommended packages: 2026-03-21T06:56:11.620 INFO:teuthology.orchestra.run.vm07.stdout: btrfs-tools 2026-03-21T06:56:11.620 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T06:56:11.620 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T06:56:11.627 INFO:teuthology.orchestra.run.vm02.stdout:The following NEW packages will be installed: 2026-03-21T06:56:11.627 INFO:teuthology.orchestra.run.vm02.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-21T06:56:11.627 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-21T06:56:11.627 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-21T06:56:11.627 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-daemon libcephfs-dev libcephfs-proxy2 libcephfs2 2026-03-21T06:56:11.627 INFO:teuthology.orchestra.run.vm02.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T06:56:11.627 INFO:teuthology.orchestra.run.vm02.stdout: libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: libradosstriper1 librdkafka1 librgw2 libsqlite3-mod-ceph libthrift-0.16.0 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: python3-pluggy python3-portend python3-prettytable python3-psutil python3-py 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: python3-pygments python3-pytest python3-rados python3-rbd python3-repoze.lru 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-rgw python3-routes python3-rsa 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: python3-simplejson python3-sklearn python3-sklearn-lib python3-tempora 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: python3-threadpoolctl python3-toml python3-wcwidth python3-webob 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse 2026-03-21T06:56:11.628 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T06:56:11.629 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be upgraded: 2026-03-21T06:56:11.629 INFO:teuthology.orchestra.run.vm02.stdout: librados2 librbd1 2026-03-21T06:56:11.660 INFO:teuthology.orchestra.run.vm07.stdout:The following NEW packages will be installed: 2026-03-21T06:56:11.660 INFO:teuthology.orchestra.run.vm07.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-21T06:56:11.660 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: libcephfs-daemon libcephfs-dev libcephfs-proxy2 libcephfs2 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: libradosstriper1 librdkafka1 librgw2 libsqlite3-mod-ceph libthrift-0.16.0 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: python3-pluggy python3-portend python3-prettytable python3-psutil python3-py 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: python3-pygments python3-pytest python3-rados python3-rbd python3-repoze.lru 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-rgw python3-routes python3-rsa 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: python3-simplejson python3-sklearn python3-sklearn-lib python3-tempora 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: python3-threadpoolctl python3-toml python3-wcwidth python3-webob 2026-03-21T06:56:11.661 INFO:teuthology.orchestra.run.vm07.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse 2026-03-21T06:56:11.662 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T06:56:11.662 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be upgraded: 2026-03-21T06:56:11.662 INFO:teuthology.orchestra.run.vm07.stdout: librados2 librbd1 2026-03-21T06:56:11.717 INFO:teuthology.orchestra.run.vm02.stdout:2 upgraded, 85 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T06:56:11.717 INFO:teuthology.orchestra.run.vm02.stdout:Need to get 281 MB of archives. 2026-03-21T06:56:11.717 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 1092 MB of additional disk space will be used. 2026-03-21T06:56:11.717 INFO:teuthology.orchestra.run.vm02.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-21T06:56:11.743 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T06:56:11.743 INFO:teuthology.orchestra.run.vm04.stdout: kpartx libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T06:56:11.743 INFO:teuthology.orchestra.run.vm04.stdout: libsgutils2-2 sg3-utils sg3-utils-udev 2026-03-21T06:56:11.743 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout:The following additional packages will be installed: 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base ceph-mgr-cephadm ceph-mgr-dashboard ceph-mgr-diskprediction-local 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-k8sevents ceph-mgr-modules-core ceph-mon ceph-osd jq 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-daemon libcephfs-proxy2 libdouble-conversion3 libfuse2 libjq1 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: liblttng-ust1 libnbd0 liboath0 libonig5 libpcre2-16-0 libqt5core5a 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: libqt5dbus5 libqt5network5 libradosstriper1 librdkafka1 librgw2 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: libsqlite3-mod-ceph libthrift-0.16.0 nvme-cli python-asyncssh-doc 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: python3-iniconfig python3-jaraco.classes python3-jaraco.collections 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.functools python3-jaraco.text python3-joblib 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: python3-kubernetes python3-natsort python3-pluggy python3-portend 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable python3-psutil python3-py python3-pygments 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: python3-pytest python3-repoze.lru python3-requests-oauthlib python3-routes 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: python3-rsa python3-simplejson python3-sklearn python3-sklearn-lib 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: python3-tempora python3-threadpoolctl python3-toml python3-wcwidth 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: python3-webob python3-websocket python3-zc.lockfile qttranslations5-l10n 2026-03-21T06:56:11.744 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T06:56:11.745 INFO:teuthology.orchestra.run.vm04.stdout:Suggested packages: 2026-03-21T06:56:11.745 INFO:teuthology.orchestra.run.vm04.stdout: python3-influxdb liblua5.3-dev luarocks python-natsort-doc python-psutil-doc 2026-03-21T06:56:11.745 INFO:teuthology.orchestra.run.vm04.stdout: subversion python-pygments-doc ttf-bitstream-vera python3-paste python3-dap 2026-03-21T06:56:11.745 INFO:teuthology.orchestra.run.vm04.stdout: python-sklearn-doc ipython3 python-webob-doc gsmartcontrol smart-notifier 2026-03-21T06:56:11.745 INFO:teuthology.orchestra.run.vm04.stdout: mailx | mailutils 2026-03-21T06:56:11.745 INFO:teuthology.orchestra.run.vm04.stdout:Recommended packages: 2026-03-21T06:56:11.745 INFO:teuthology.orchestra.run.vm04.stdout: btrfs-tools 2026-03-21T06:56:11.783 INFO:teuthology.orchestra.run.vm04.stdout:The following NEW packages will be installed: 2026-03-21T06:56:11.783 INFO:teuthology.orchestra.run.vm04.stdout: ceph ceph-base ceph-common ceph-fuse ceph-mds ceph-mgr ceph-mgr-cephadm 2026-03-21T06:56:11.783 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-dashboard ceph-mgr-diskprediction-local ceph-mgr-k8sevents 2026-03-21T06:56:11.783 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core ceph-mon ceph-osd ceph-test ceph-volume cephadm jq 2026-03-21T06:56:11.783 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-daemon libcephfs-dev libcephfs-proxy2 libcephfs2 2026-03-21T06:56:11.783 INFO:teuthology.orchestra.run.vm04.stdout: libdouble-conversion3 libfuse2 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T06:56:11.783 INFO:teuthology.orchestra.run.vm04.stdout: libonig5 libpcre2-16-0 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: libradosstriper1 librdkafka1 librgw2 libsqlite3-mod-ceph libthrift-0.16.0 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-argparse python3-ceph-common python3-cephfs python3-cheroot 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy3 python3-google-auth python3-iniconfig 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: python3-pluggy python3-portend python3-prettytable python3-psutil python3-py 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: python3-pygments python3-pytest python3-rados python3-rbd python3-repoze.lru 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-rgw python3-routes python3-rsa 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: python3-simplejson python3-sklearn python3-sklearn-lib python3-tempora 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: python3-threadpoolctl python3-toml python3-wcwidth python3-webob 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket python3-zc.lockfile qttranslations5-l10n radosgw rbd-fuse 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be upgraded: 2026-03-21T06:56:11.784 INFO:teuthology.orchestra.run.vm04.stdout: librados2 librbd1 2026-03-21T06:56:11.813 INFO:teuthology.orchestra.run.vm04.stdout:2 upgraded, 85 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T06:56:11.813 INFO:teuthology.orchestra.run.vm04.stdout:Need to get 281 MB of archives. 2026-03-21T06:56:11.813 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 1092 MB of additional disk space will be used. 2026-03-21T06:56:11.813 INFO:teuthology.orchestra.run.vm04.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-21T06:56:11.851 INFO:teuthology.orchestra.run.vm04.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-21T06:56:11.853 INFO:teuthology.orchestra.run.vm04.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-21T06:56:11.860 INFO:teuthology.orchestra.run.vm04.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-21T06:56:11.879 INFO:teuthology.orchestra.run.vm07.stdout:2 upgraded, 85 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T06:56:11.879 INFO:teuthology.orchestra.run.vm07.stdout:Need to get 281 MB of archives. 2026-03-21T06:56:11.879 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 1092 MB of additional disk space will be used. 2026-03-21T06:56:11.879 INFO:teuthology.orchestra.run.vm07.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 liblttng-ust1 amd64 2.13.1-1ubuntu1 [190 kB] 2026-03-21T06:56:11.886 INFO:teuthology.orchestra.run.vm04.stdout:Get:5 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-21T06:56:11.888 INFO:teuthology.orchestra.run.vm04.stdout:Get:6 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-21T06:56:11.894 INFO:teuthology.orchestra.run.vm04.stdout:Get:7 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-21T06:56:11.897 INFO:teuthology.orchestra.run.vm02.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-21T06:56:11.899 INFO:teuthology.orchestra.run.vm02.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-21T06:56:11.904 INFO:teuthology.orchestra.run.vm04.stdout:Get:8 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-21T06:56:11.904 INFO:teuthology.orchestra.run.vm04.stdout:Get:9 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-21T06:56:11.905 INFO:teuthology.orchestra.run.vm04.stdout:Get:10 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-21T06:56:11.905 INFO:teuthology.orchestra.run.vm04.stdout:Get:11 http://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-21T06:56:11.913 INFO:teuthology.orchestra.run.vm04.stdout:Get:12 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-21T06:56:11.913 INFO:teuthology.orchestra.run.vm04.stdout:Get:13 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-21T06:56:11.913 INFO:teuthology.orchestra.run.vm04.stdout:Get:14 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-21T06:56:11.914 INFO:teuthology.orchestra.run.vm04.stdout:Get:15 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-21T06:56:11.914 INFO:teuthology.orchestra.run.vm04.stdout:Get:16 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-21T06:56:11.914 INFO:teuthology.orchestra.run.vm04.stdout:Get:17 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-21T06:56:11.915 INFO:teuthology.orchestra.run.vm04.stdout:Get:18 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-21T06:56:11.915 INFO:teuthology.orchestra.run.vm04.stdout:Get:19 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-21T06:56:11.921 INFO:teuthology.orchestra.run.vm04.stdout:Get:20 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-21T06:56:11.921 INFO:teuthology.orchestra.run.vm04.stdout:Get:21 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-21T06:56:11.929 INFO:teuthology.orchestra.run.vm04.stdout:Get:22 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-21T06:56:11.929 INFO:teuthology.orchestra.run.vm04.stdout:Get:23 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-21T06:56:11.930 INFO:teuthology.orchestra.run.vm04.stdout:Get:24 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-21T06:56:11.931 INFO:teuthology.orchestra.run.vm04.stdout:Get:25 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-21T06:56:11.932 INFO:teuthology.orchestra.run.vm04.stdout:Get:26 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-21T06:56:11.932 INFO:teuthology.orchestra.run.vm04.stdout:Get:27 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-21T06:56:11.957 INFO:teuthology.orchestra.run.vm04.stdout:Get:28 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-21T06:56:11.958 INFO:teuthology.orchestra.run.vm04.stdout:Get:29 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-21T06:56:11.958 INFO:teuthology.orchestra.run.vm04.stdout:Get:30 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-21T06:56:11.978 INFO:teuthology.orchestra.run.vm04.stdout:Get:31 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-21T06:56:11.978 INFO:teuthology.orchestra.run.vm04.stdout:Get:32 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-21T06:56:11.979 INFO:teuthology.orchestra.run.vm04.stdout:Get:33 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-21T06:56:11.979 INFO:teuthology.orchestra.run.vm04.stdout:Get:34 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-21T06:56:11.979 INFO:teuthology.orchestra.run.vm04.stdout:Get:35 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-21T06:56:11.980 INFO:teuthology.orchestra.run.vm04.stdout:Get:36 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-21T06:56:11.981 INFO:teuthology.orchestra.run.vm04.stdout:Get:37 http://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-21T06:56:11.982 INFO:teuthology.orchestra.run.vm04.stdout:Get:38 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-21T06:56:11.982 INFO:teuthology.orchestra.run.vm02.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-21T06:56:11.983 INFO:teuthology.orchestra.run.vm04.stdout:Get:39 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-21T06:56:11.988 INFO:teuthology.orchestra.run.vm04.stdout:Get:40 http://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-21T06:56:11.991 INFO:teuthology.orchestra.run.vm04.stdout:Get:41 http://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-21T06:56:11.993 INFO:teuthology.orchestra.run.vm04.stdout:Get:42 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-21T06:56:12.006 INFO:teuthology.orchestra.run.vm04.stdout:Get:43 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-21T06:56:12.009 INFO:teuthology.orchestra.run.vm04.stdout:Get:44 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-21T06:56:12.009 INFO:teuthology.orchestra.run.vm04.stdout:Get:45 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-21T06:56:12.009 INFO:teuthology.orchestra.run.vm04.stdout:Get:46 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-21T06:56:12.010 INFO:teuthology.orchestra.run.vm04.stdout:Get:47 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-21T06:56:12.011 INFO:teuthology.orchestra.run.vm04.stdout:Get:48 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-21T06:56:12.017 INFO:teuthology.orchestra.run.vm04.stdout:Get:49 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-21T06:56:12.017 INFO:teuthology.orchestra.run.vm04.stdout:Get:50 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-21T06:56:12.018 INFO:teuthology.orchestra.run.vm04.stdout:Get:51 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-21T06:56:12.018 INFO:teuthology.orchestra.run.vm04.stdout:Get:52 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-21T06:56:12.023 INFO:teuthology.orchestra.run.vm04.stdout:Get:53 http://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-21T06:56:12.043 INFO:teuthology.orchestra.run.vm04.stdout:Get:54 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-21T06:56:12.094 INFO:teuthology.orchestra.run.vm02.stdout:Get:5 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-21T06:56:12.100 INFO:teuthology.orchestra.run.vm02.stdout:Get:6 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-21T06:56:12.117 INFO:teuthology.orchestra.run.vm02.stdout:Get:7 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-21T06:56:12.122 INFO:teuthology.orchestra.run.vm02.stdout:Get:8 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-21T06:56:12.124 INFO:teuthology.orchestra.run.vm02.stdout:Get:9 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-21T06:56:12.124 INFO:teuthology.orchestra.run.vm02.stdout:Get:10 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-21T06:56:12.125 INFO:teuthology.orchestra.run.vm02.stdout:Get:11 http://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-21T06:56:12.134 INFO:teuthology.orchestra.run.vm02.stdout:Get:12 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-21T06:56:12.135 INFO:teuthology.orchestra.run.vm02.stdout:Get:13 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-21T06:56:12.135 INFO:teuthology.orchestra.run.vm02.stdout:Get:14 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-21T06:56:12.167 INFO:teuthology.orchestra.run.vm02.stdout:Get:15 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-21T06:56:12.167 INFO:teuthology.orchestra.run.vm02.stdout:Get:16 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-21T06:56:12.168 INFO:teuthology.orchestra.run.vm02.stdout:Get:17 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-21T06:56:12.168 INFO:teuthology.orchestra.run.vm02.stdout:Get:18 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-21T06:56:12.168 INFO:teuthology.orchestra.run.vm02.stdout:Get:19 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-21T06:56:12.168 INFO:teuthology.orchestra.run.vm02.stdout:Get:20 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-21T06:56:12.203 INFO:teuthology.orchestra.run.vm02.stdout:Get:21 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-21T06:56:12.205 INFO:teuthology.orchestra.run.vm02.stdout:Get:22 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-21T06:56:12.205 INFO:teuthology.orchestra.run.vm02.stdout:Get:23 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-21T06:56:12.206 INFO:teuthology.orchestra.run.vm02.stdout:Get:24 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-21T06:56:12.239 INFO:teuthology.orchestra.run.vm02.stdout:Get:25 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-21T06:56:12.239 INFO:teuthology.orchestra.run.vm02.stdout:Get:26 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-21T06:56:12.240 INFO:teuthology.orchestra.run.vm02.stdout:Get:27 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-21T06:56:12.258 INFO:teuthology.orchestra.run.vm02.stdout:Get:28 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-21T06:56:12.259 INFO:teuthology.orchestra.run.vm02.stdout:Get:29 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-21T06:56:12.259 INFO:teuthology.orchestra.run.vm02.stdout:Get:30 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-21T06:56:12.269 INFO:teuthology.orchestra.run.vm07.stdout:Get:2 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 20.2.0-712-g70f8415b-1jammy [2867 kB] 2026-03-21T06:56:12.321 INFO:teuthology.orchestra.run.vm02.stdout:Get:31 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-21T06:56:12.321 INFO:teuthology.orchestra.run.vm02.stdout:Get:32 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-21T06:56:12.321 INFO:teuthology.orchestra.run.vm02.stdout:Get:33 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-21T06:56:12.321 INFO:teuthology.orchestra.run.vm02.stdout:Get:34 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-21T06:56:12.322 INFO:teuthology.orchestra.run.vm02.stdout:Get:35 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-21T06:56:12.322 INFO:teuthology.orchestra.run.vm02.stdout:Get:36 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-21T06:56:12.323 INFO:teuthology.orchestra.run.vm02.stdout:Get:37 http://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-21T06:56:12.324 INFO:teuthology.orchestra.run.vm02.stdout:Get:38 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-21T06:56:12.325 INFO:teuthology.orchestra.run.vm02.stdout:Get:39 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-21T06:56:12.357 INFO:teuthology.orchestra.run.vm02.stdout:Get:40 http://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-21T06:56:12.360 INFO:teuthology.orchestra.run.vm02.stdout:Get:41 http://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-21T06:56:12.383 INFO:teuthology.orchestra.run.vm04.stdout:Get:55 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 20.2.0-712-g70f8415b-1jammy [2867 kB] 2026-03-21T06:56:12.392 INFO:teuthology.orchestra.run.vm02.stdout:Get:42 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-21T06:56:12.396 INFO:teuthology.orchestra.run.vm02.stdout:Get:43 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-21T06:56:12.399 INFO:teuthology.orchestra.run.vm02.stdout:Get:44 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-21T06:56:12.399 INFO:teuthology.orchestra.run.vm02.stdout:Get:45 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-21T06:56:12.399 INFO:teuthology.orchestra.run.vm02.stdout:Get:46 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-21T06:56:12.400 INFO:teuthology.orchestra.run.vm02.stdout:Get:47 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-21T06:56:12.401 INFO:teuthology.orchestra.run.vm02.stdout:Get:48 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-21T06:56:12.405 INFO:teuthology.orchestra.run.vm07.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libdouble-conversion3 amd64 3.1.7-4 [39.0 kB] 2026-03-21T06:56:12.418 INFO:teuthology.orchestra.run.vm07.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libpcre2-16-0 amd64 10.39-3ubuntu0.1 [203 kB] 2026-03-21T06:56:12.464 INFO:teuthology.orchestra.run.vm02.stdout:Get:49 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-21T06:56:12.464 INFO:teuthology.orchestra.run.vm02.stdout:Get:50 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-21T06:56:12.465 INFO:teuthology.orchestra.run.vm02.stdout:Get:51 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-21T06:56:12.465 INFO:teuthology.orchestra.run.vm02.stdout:Get:52 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-21T06:56:12.466 INFO:teuthology.orchestra.run.vm02.stdout:Get:53 http://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-21T06:56:12.524 INFO:teuthology.orchestra.run.vm07.stdout:Get:5 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5core5a amd64 5.15.3+dfsg-2ubuntu0.2 [2006 kB] 2026-03-21T06:56:12.573 INFO:teuthology.orchestra.run.vm02.stdout:Get:54 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-21T06:56:12.751 INFO:teuthology.orchestra.run.vm02.stdout:Get:55 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librbd1 amd64 20.2.0-712-g70f8415b-1jammy [2867 kB] 2026-03-21T06:56:12.821 INFO:teuthology.orchestra.run.vm07.stdout:Get:6 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5dbus5 amd64 5.15.3+dfsg-2ubuntu0.2 [222 kB] 2026-03-21T06:56:12.835 INFO:teuthology.orchestra.run.vm07.stdout:Get:7 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 libqt5network5 amd64 5.15.3+dfsg-2ubuntu0.2 [731 kB] 2026-03-21T06:56:12.874 INFO:teuthology.orchestra.run.vm07.stdout:Get:8 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libthrift-0.16.0 amd64 0.16.0-2 [267 kB] 2026-03-21T06:56:12.886 INFO:teuthology.orchestra.run.vm07.stdout:Get:9 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libnbd0 amd64 1.10.5-1 [71.3 kB] 2026-03-21T06:56:12.888 INFO:teuthology.orchestra.run.vm07.stdout:Get:10 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-wcwidth all 0.2.5+dfsg1-1 [21.9 kB] 2026-03-21T06:56:12.889 INFO:teuthology.orchestra.run.vm07.stdout:Get:11 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-prettytable all 2.5.0-2 [31.3 kB] 2026-03-21T06:56:12.891 INFO:teuthology.orchestra.run.vm07.stdout:Get:12 http://archive.ubuntu.com/ubuntu jammy/universe amd64 librdkafka1 amd64 1.8.0-1build1 [633 kB] 2026-03-21T06:56:12.915 INFO:teuthology.orchestra.run.vm07.stdout:Get:13 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 liboath0 amd64 2.6.7-3ubuntu0.1 [41.3 kB] 2026-03-21T06:56:12.916 INFO:teuthology.orchestra.run.vm07.stdout:Get:14 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.functools all 3.4.0-2 [9030 B] 2026-03-21T06:56:12.924 INFO:teuthology.orchestra.run.vm07.stdout:Get:15 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-cheroot all 8.5.2+ds1-1ubuntu3.1 [71.1 kB] 2026-03-21T06:56:13.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:12 vm04 bash[20194]: cluster 2026-03-21T06:56:11.606800+0000 mgr.x (mgr.14152) 367 : cluster [DBG] pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.8 KiB/s rd, 2 op/s 2026-03-21T06:56:13.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:12 vm04 bash[20194]: cluster 2026-03-21T06:56:11.606800+0000 mgr.x (mgr.14152) 367 : cluster [DBG] pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.8 KiB/s rd, 2 op/s 2026-03-21T06:56:13.020 INFO:teuthology.orchestra.run.vm07.stdout:Get:16 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.classes all 3.2.1-3 [6452 B] 2026-03-21T06:56:13.020 INFO:teuthology.orchestra.run.vm07.stdout:Get:17 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.text all 3.6.0-2 [8716 B] 2026-03-21T06:56:13.021 INFO:teuthology.orchestra.run.vm07.stdout:Get:18 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jaraco.collections all 3.4.0-2 [11.4 kB] 2026-03-21T06:56:13.021 INFO:teuthology.orchestra.run.vm07.stdout:Get:19 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-tempora all 4.1.2-1 [14.8 kB] 2026-03-21T06:56:13.021 INFO:teuthology.orchestra.run.vm07.stdout:Get:20 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-portend all 3.0.0-1 [7240 B] 2026-03-21T06:56:13.021 INFO:teuthology.orchestra.run.vm07.stdout:Get:21 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-zc.lockfile all 2.0-1 [8980 B] 2026-03-21T06:56:13.022 INFO:teuthology.orchestra.run.vm07.stdout:Get:22 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cherrypy3 all 18.6.1-4 [208 kB] 2026-03-21T06:56:13.024 INFO:teuthology.orchestra.run.vm07.stdout:Get:23 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-natsort all 8.0.2-1 [35.3 kB] 2026-03-21T06:56:13.028 INFO:teuthology.orchestra.run.vm07.stdout:Get:24 http://archive.ubuntu.com/ubuntu jammy/universe amd64 libfuse2 amd64 2.9.9-5ubuntu3 [90.3 kB] 2026-03-21T06:56:13.129 INFO:teuthology.orchestra.run.vm07.stdout:Get:25 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python3-asyncssh all 2.5.0-1ubuntu0.1 [189 kB] 2026-03-21T06:56:13.132 INFO:teuthology.orchestra.run.vm07.stdout:Get:26 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-repoze.lru all 0.7-2 [12.1 kB] 2026-03-21T06:56:13.132 INFO:teuthology.orchestra.run.vm07.stdout:Get:27 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-routes all 2.5.1-1ubuntu1 [89.0 kB] 2026-03-21T06:56:13.133 INFO:teuthology.orchestra.run.vm07.stdout:Get:28 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn-lib amd64 0.23.2-5ubuntu6 [2058 kB] 2026-03-21T06:56:13.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:12 vm02 bash[17657]: cluster 2026-03-21T06:56:11.606800+0000 mgr.x (mgr.14152) 367 : cluster [DBG] pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.8 KiB/s rd, 2 op/s 2026-03-21T06:56:13.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:12 vm02 bash[17657]: cluster 2026-03-21T06:56:11.606800+0000 mgr.x (mgr.14152) 367 : cluster [DBG] pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.8 KiB/s rd, 2 op/s 2026-03-21T06:56:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:12 vm07 bash[19945]: cluster 2026-03-21T06:56:11.606800+0000 mgr.x (mgr.14152) 367 : cluster [DBG] pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.8 KiB/s rd, 2 op/s 2026-03-21T06:56:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:12 vm07 bash[19945]: cluster 2026-03-21T06:56:11.606800+0000 mgr.x (mgr.14152) 367 : cluster [DBG] pgmap v300: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.8 KiB/s rd, 2 op/s 2026-03-21T06:56:13.169 INFO:teuthology.orchestra.run.vm07.stdout:Get:29 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-joblib all 0.17.0-4ubuntu1 [204 kB] 2026-03-21T06:56:13.172 INFO:teuthology.orchestra.run.vm07.stdout:Get:30 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-threadpoolctl all 3.1.0-1 [21.3 kB] 2026-03-21T06:56:13.173 INFO:teuthology.orchestra.run.vm07.stdout:Get:31 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-sklearn all 0.23.2-5ubuntu6 [1829 kB] 2026-03-21T06:56:13.197 INFO:teuthology.orchestra.run.vm04.stdout:Get:56 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 20.2.0-712-g70f8415b-1jammy [3583 kB] 2026-03-21T06:56:13.254 INFO:teuthology.orchestra.run.vm07.stdout:Get:32 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-cachetools all 5.0.0-1 [9722 B] 2026-03-21T06:56:13.255 INFO:teuthology.orchestra.run.vm07.stdout:Get:33 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-rsa all 4.8-1 [28.4 kB] 2026-03-21T06:56:13.255 INFO:teuthology.orchestra.run.vm07.stdout:Get:34 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-google-auth all 1.5.1-3 [35.7 kB] 2026-03-21T06:56:13.255 INFO:teuthology.orchestra.run.vm07.stdout:Get:35 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-requests-oauthlib all 1.3.0+ds-0.1 [18.7 kB] 2026-03-21T06:56:13.256 INFO:teuthology.orchestra.run.vm07.stdout:Get:36 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-websocket all 1.2.3-1 [34.7 kB] 2026-03-21T06:56:13.256 INFO:teuthology.orchestra.run.vm07.stdout:Get:37 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-kubernetes all 12.0.1-1ubuntu1 [353 kB] 2026-03-21T06:56:13.320 INFO:teuthology.orchestra.run.vm04.stdout:Get:57 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 20.2.0-712-g70f8415b-1jammy [829 kB] 2026-03-21T06:56:13.330 INFO:teuthology.orchestra.run.vm04.stdout:Get:58 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 20.2.0-712-g70f8415b-1jammy [364 kB] 2026-03-21T06:56:13.335 INFO:teuthology.orchestra.run.vm04.stdout:Get:59 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 20.2.0-712-g70f8415b-1jammy [32.8 kB] 2026-03-21T06:56:13.336 INFO:teuthology.orchestra.run.vm04.stdout:Get:60 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 20.2.0-712-g70f8415b-1jammy [184 kB] 2026-03-21T06:56:13.339 INFO:teuthology.orchestra.run.vm04.stdout:Get:61 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 20.2.0-712-g70f8415b-1jammy [83.8 kB] 2026-03-21T06:56:13.340 INFO:teuthology.orchestra.run.vm07.stdout:Get:38 http://archive.ubuntu.com/ubuntu jammy/main amd64 libonig5 amd64 6.9.7.1-2build1 [172 kB] 2026-03-21T06:56:13.340 INFO:teuthology.orchestra.run.vm04.stdout:Get:62 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 20.2.0-712-g70f8415b-1jammy [341 kB] 2026-03-21T06:56:13.343 INFO:teuthology.orchestra.run.vm07.stdout:Get:39 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 libjq1 amd64 1.6-2.1ubuntu3.1 [133 kB] 2026-03-21T06:56:13.344 INFO:teuthology.orchestra.run.vm07.stdout:Get:40 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 jq amd64 1.6-2.1ubuntu3.1 [52.5 kB] 2026-03-21T06:56:13.346 INFO:teuthology.orchestra.run.vm04.stdout:Get:63 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 20.2.0-712-g70f8415b-1jammy [8697 kB] 2026-03-21T06:56:13.445 INFO:teuthology.orchestra.run.vm07.stdout:Get:41 http://archive.ubuntu.com/ubuntu jammy/main amd64 socat amd64 1.7.4.1-3ubuntu4 [349 kB] 2026-03-21T06:56:13.450 INFO:teuthology.orchestra.run.vm07.stdout:Get:42 http://archive.ubuntu.com/ubuntu jammy/universe amd64 xmlstarlet amd64 1.6.1-2.1 [265 kB] 2026-03-21T06:56:13.454 INFO:teuthology.orchestra.run.vm07.stdout:Get:43 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 nvme-cli amd64 1.16-3ubuntu0.3 [474 kB] 2026-03-21T06:56:13.461 INFO:teuthology.orchestra.run.vm07.stdout:Get:44 http://archive.ubuntu.com/ubuntu jammy-updates/universe amd64 python-asyncssh-doc all 2.5.0-1ubuntu0.1 [309 kB] 2026-03-21T06:56:13.465 INFO:teuthology.orchestra.run.vm07.stdout:Get:45 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-iniconfig all 1.1.1-2 [6024 B] 2026-03-21T06:56:13.465 INFO:teuthology.orchestra.run.vm07.stdout:Get:46 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pluggy all 0.13.0-7.1 [19.0 kB] 2026-03-21T06:56:13.465 INFO:teuthology.orchestra.run.vm07.stdout:Get:47 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-psutil amd64 5.9.0-1build1 [158 kB] 2026-03-21T06:56:13.548 INFO:teuthology.orchestra.run.vm07.stdout:Get:48 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-py all 1.10.0-1 [71.9 kB] 2026-03-21T06:56:13.549 INFO:teuthology.orchestra.run.vm07.stdout:Get:49 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-pygments all 2.11.2+dfsg-2ubuntu0.1 [750 kB] 2026-03-21T06:56:13.558 INFO:teuthology.orchestra.run.vm07.stdout:Get:50 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-toml all 0.10.2-1 [16.5 kB] 2026-03-21T06:56:13.652 INFO:teuthology.orchestra.run.vm07.stdout:Get:51 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-pytest all 6.2.5-1ubuntu2 [214 kB] 2026-03-21T06:56:13.654 INFO:teuthology.orchestra.run.vm07.stdout:Get:52 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-simplejson amd64 3.17.6-1build1 [54.7 kB] 2026-03-21T06:56:13.655 INFO:teuthology.orchestra.run.vm07.stdout:Get:53 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 python3-webob all 1:1.8.6-1.1ubuntu0.1 [86.7 kB] 2026-03-21T06:56:13.656 INFO:teuthology.orchestra.run.vm07.stdout:Get:54 http://archive.ubuntu.com/ubuntu jammy/universe amd64 qttranslations5-l10n all 5.15.3-1 [1983 kB] 2026-03-21T06:56:13.825 INFO:teuthology.orchestra.run.vm07.stdout:Get:55 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 smartmontools amd64 7.2-1ubuntu0.1 [583 kB] 2026-03-21T06:56:13.892 INFO:teuthology.orchestra.run.vm04.stdout:Get:64 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 20.2.0-712-g70f8415b-1jammy [112 kB] 2026-03-21T06:56:13.892 INFO:teuthology.orchestra.run.vm04.stdout:Get:65 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 20.2.0-712-g70f8415b-1jammy [261 kB] 2026-03-21T06:56:13.893 INFO:teuthology.orchestra.run.vm04.stdout:Get:66 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 20.2.0-712-g70f8415b-1jammy [29.3 MB] 2026-03-21T06:56:14.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:13 vm02 bash[17657]: cluster 2026-03-21T06:56:13.607122+0000 mgr.x (mgr.14152) 368 : cluster [DBG] pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-21T06:56:14.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:13 vm02 bash[17657]: cluster 2026-03-21T06:56:13.607122+0000 mgr.x (mgr.14152) 368 : cluster [DBG] pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-21T06:56:14.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:13 vm07 bash[19945]: cluster 2026-03-21T06:56:13.607122+0000 mgr.x (mgr.14152) 368 : cluster [DBG] pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-21T06:56:14.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:13 vm07 bash[19945]: cluster 2026-03-21T06:56:13.607122+0000 mgr.x (mgr.14152) 368 : cluster [DBG] pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-21T06:56:14.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:13 vm04 bash[20194]: cluster 2026-03-21T06:56:13.607122+0000 mgr.x (mgr.14152) 368 : cluster [DBG] pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-21T06:56:14.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:13 vm04 bash[20194]: cluster 2026-03-21T06:56:13.607122+0000 mgr.x (mgr.14152) 368 : cluster [DBG] pgmap v301: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.2 KiB/s rd, 1 op/s 2026-03-21T06:56:14.877 INFO:teuthology.orchestra.run.vm07.stdout:Get:56 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 20.2.0-712-g70f8415b-1jammy [3583 kB] 2026-03-21T06:56:15.203 INFO:teuthology.orchestra.run.vm04.stdout:Get:67 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 20.2.0-712-g70f8415b-1jammy [5415 kB] 2026-03-21T06:56:15.352 INFO:teuthology.orchestra.run.vm04.stdout:Get:68 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 20.2.0-712-g70f8415b-1jammy [246 kB] 2026-03-21T06:56:15.433 INFO:teuthology.orchestra.run.vm04.stdout:Get:69 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 20.2.0-712-g70f8415b-1jammy [124 kB] 2026-03-21T06:56:15.433 INFO:teuthology.orchestra.run.vm04.stdout:Get:70 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 20.2.0-712-g70f8415b-1jammy [906 kB] 2026-03-21T06:56:15.441 INFO:teuthology.orchestra.run.vm04.stdout:Get:71 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 20.2.0-712-g70f8415b-1jammy [6399 kB] 2026-03-21T06:56:15.681 INFO:teuthology.orchestra.run.vm04.stdout:Get:72 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 20.2.0-712-g70f8415b-1jammy [21.7 MB] 2026-03-21T06:56:16.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:15 vm02 bash[17657]: cluster 2026-03-21T06:56:15.607476+0000 mgr.x (mgr.14152) 369 : cluster [DBG] pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 2 op/s 2026-03-21T06:56:16.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:15 vm02 bash[17657]: cluster 2026-03-21T06:56:15.607476+0000 mgr.x (mgr.14152) 369 : cluster [DBG] pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 2 op/s 2026-03-21T06:56:16.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:15 vm07 bash[19945]: cluster 2026-03-21T06:56:15.607476+0000 mgr.x (mgr.14152) 369 : cluster [DBG] pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 2 op/s 2026-03-21T06:56:16.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:15 vm07 bash[19945]: cluster 2026-03-21T06:56:15.607476+0000 mgr.x (mgr.14152) 369 : cluster [DBG] pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 2 op/s 2026-03-21T06:56:16.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:15 vm04 bash[20194]: cluster 2026-03-21T06:56:15.607476+0000 mgr.x (mgr.14152) 369 : cluster [DBG] pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 2 op/s 2026-03-21T06:56:16.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:15 vm04 bash[20194]: cluster 2026-03-21T06:56:15.607476+0000 mgr.x (mgr.14152) 369 : cluster [DBG] pgmap v302: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.0 KiB/s rd, 2 op/s 2026-03-21T06:56:16.514 INFO:teuthology.orchestra.run.vm04.stdout:Get:73 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 20.2.0-712-g70f8415b-1jammy [14.1 kB] 2026-03-21T06:56:16.514 INFO:teuthology.orchestra.run.vm04.stdout:Get:74 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 20.2.0-712-g70f8415b-1jammy [955 kB] 2026-03-21T06:56:16.519 INFO:teuthology.orchestra.run.vm04.stdout:Get:75 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 20.2.0-712-g70f8415b-1jammy [2341 kB] 2026-03-21T06:56:16.632 INFO:teuthology.orchestra.run.vm04.stdout:Get:76 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 20.2.0-712-g70f8415b-1jammy [1049 kB] 2026-03-21T06:56:16.641 INFO:teuthology.orchestra.run.vm04.stdout:Get:77 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 20.2.0-712-g70f8415b-1jammy [179 kB] 2026-03-21T06:56:16.644 INFO:teuthology.orchestra.run.vm04.stdout:Get:78 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 20.2.0-712-g70f8415b-1jammy [45.5 MB] 2026-03-21T06:56:16.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:16 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:56:16.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:16 vm02 bash[17657]: audit 2026-03-21T06:56:16.633937+0000 mgr.x (mgr.14152) 370 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:16.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:16 vm02 bash[17657]: audit 2026-03-21T06:56:16.633937+0000 mgr.x (mgr.14152) 370 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:16 vm07 bash[19945]: audit 2026-03-21T06:56:16.633937+0000 mgr.x (mgr.14152) 370 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:16 vm07 bash[19945]: audit 2026-03-21T06:56:16.633937+0000 mgr.x (mgr.14152) 370 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:17.239 INFO:teuthology.orchestra.run.vm07.stdout:Get:57 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 20.2.0-712-g70f8415b-1jammy [829 kB] 2026-03-21T06:56:17.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:16 vm04 bash[20194]: audit 2026-03-21T06:56:16.633937+0000 mgr.x (mgr.14152) 370 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:17.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:16 vm04 bash[20194]: audit 2026-03-21T06:56:16.633937+0000 mgr.x (mgr.14152) 370 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:17.794 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:17 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:56:17.822 INFO:teuthology.orchestra.run.vm07.stdout:Get:58 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 20.2.0-712-g70f8415b-1jammy [364 kB] 2026-03-21T06:56:18.056 INFO:teuthology.orchestra.run.vm07.stdout:Get:59 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 20.2.0-712-g70f8415b-1jammy [32.8 kB] 2026-03-21T06:56:18.056 INFO:teuthology.orchestra.run.vm07.stdout:Get:60 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 20.2.0-712-g70f8415b-1jammy [184 kB] 2026-03-21T06:56:18.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:17 vm02 bash[17657]: audit 2026-03-21T06:56:17.470595+0000 mgr.x (mgr.14152) 371 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:18.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:17 vm02 bash[17657]: audit 2026-03-21T06:56:17.470595+0000 mgr.x (mgr.14152) 371 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:18.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:17 vm02 bash[17657]: cluster 2026-03-21T06:56:17.607806+0000 mgr.x (mgr.14152) 372 : cluster [DBG] pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 1 op/s 2026-03-21T06:56:18.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:17 vm02 bash[17657]: cluster 2026-03-21T06:56:17.607806+0000 mgr.x (mgr.14152) 372 : cluster [DBG] pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 1 op/s 2026-03-21T06:56:18.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:17 vm07 bash[19945]: audit 2026-03-21T06:56:17.470595+0000 mgr.x (mgr.14152) 371 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:18.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:17 vm07 bash[19945]: audit 2026-03-21T06:56:17.470595+0000 mgr.x (mgr.14152) 371 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:18.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:17 vm07 bash[19945]: cluster 2026-03-21T06:56:17.607806+0000 mgr.x (mgr.14152) 372 : cluster [DBG] pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 1 op/s 2026-03-21T06:56:18.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:17 vm07 bash[19945]: cluster 2026-03-21T06:56:17.607806+0000 mgr.x (mgr.14152) 372 : cluster [DBG] pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 1 op/s 2026-03-21T06:56:18.173 INFO:teuthology.orchestra.run.vm07.stdout:Get:61 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 20.2.0-712-g70f8415b-1jammy [83.8 kB] 2026-03-21T06:56:18.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:17 vm04 bash[20194]: audit 2026-03-21T06:56:17.470595+0000 mgr.x (mgr.14152) 371 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:18.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:17 vm04 bash[20194]: audit 2026-03-21T06:56:17.470595+0000 mgr.x (mgr.14152) 371 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:18.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:17 vm04 bash[20194]: cluster 2026-03-21T06:56:17.607806+0000 mgr.x (mgr.14152) 372 : cluster [DBG] pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 1 op/s 2026-03-21T06:56:18.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:17 vm04 bash[20194]: cluster 2026-03-21T06:56:17.607806+0000 mgr.x (mgr.14152) 372 : cluster [DBG] pgmap v303: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.8 KiB/s rd, 1 op/s 2026-03-21T06:56:18.289 INFO:teuthology.orchestra.run.vm07.stdout:Get:62 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 20.2.0-712-g70f8415b-1jammy [341 kB] 2026-03-21T06:56:18.523 INFO:teuthology.orchestra.run.vm07.stdout:Get:63 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 20.2.0-712-g70f8415b-1jammy [8697 kB] 2026-03-21T06:56:18.841 INFO:teuthology.orchestra.run.vm04.stdout:Get:79 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 20.2.0-712-g70f8415b-1jammy [8625 kB] 2026-03-21T06:56:19.278 INFO:teuthology.orchestra.run.vm04.stdout:Get:80 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 20.2.0-712-g70f8415b-1jammy [14.2 kB] 2026-03-21T06:56:19.278 INFO:teuthology.orchestra.run.vm04.stdout:Get:81 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 20.2.0-712-g70f8415b-1jammy [99.5 MB] 2026-03-21T06:56:20.900 INFO:teuthology.orchestra.run.vm02.stdout:Get:56 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librados2 amd64 20.2.0-712-g70f8415b-1jammy [3583 kB] 2026-03-21T06:56:21.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:20 vm04 bash[20194]: cluster 2026-03-21T06:56:19.608145+0000 mgr.x (mgr.14152) 373 : cluster [DBG] pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:21.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:20 vm04 bash[20194]: cluster 2026-03-21T06:56:19.608145+0000 mgr.x (mgr.14152) 373 : cluster [DBG] pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:21.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:20 vm02 bash[17657]: cluster 2026-03-21T06:56:19.608145+0000 mgr.x (mgr.14152) 373 : cluster [DBG] pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:21.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:20 vm02 bash[17657]: cluster 2026-03-21T06:56:19.608145+0000 mgr.x (mgr.14152) 373 : cluster [DBG] pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:20 vm07 bash[19945]: cluster 2026-03-21T06:56:19.608145+0000 mgr.x (mgr.14152) 373 : cluster [DBG] pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:20 vm07 bash[19945]: cluster 2026-03-21T06:56:19.608145+0000 mgr.x (mgr.14152) 373 : cluster [DBG] pgmap v304: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:22.933 INFO:teuthology.orchestra.run.vm02.stdout:Get:57 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs2 amd64 20.2.0-712-g70f8415b-1jammy [829 kB] 2026-03-21T06:56:23.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:22 vm04 bash[20194]: cluster 2026-03-21T06:56:21.608521+0000 mgr.x (mgr.14152) 374 : cluster [DBG] pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:23.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:22 vm04 bash[20194]: cluster 2026-03-21T06:56:21.608521+0000 mgr.x (mgr.14152) 374 : cluster [DBG] pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:23.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:22 vm02 bash[17657]: cluster 2026-03-21T06:56:21.608521+0000 mgr.x (mgr.14152) 374 : cluster [DBG] pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:23.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:22 vm02 bash[17657]: cluster 2026-03-21T06:56:21.608521+0000 mgr.x (mgr.14152) 374 : cluster [DBG] pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:22 vm07 bash[19945]: cluster 2026-03-21T06:56:21.608521+0000 mgr.x (mgr.14152) 374 : cluster [DBG] pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:22 vm07 bash[19945]: cluster 2026-03-21T06:56:21.608521+0000 mgr.x (mgr.14152) 374 : cluster [DBG] pgmap v305: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:23.173 INFO:teuthology.orchestra.run.vm02.stdout:Get:58 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rados amd64 20.2.0-712-g70f8415b-1jammy [364 kB] 2026-03-21T06:56:23.293 INFO:teuthology.orchestra.run.vm02.stdout:Get:59 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-argparse all 20.2.0-712-g70f8415b-1jammy [32.8 kB] 2026-03-21T06:56:23.293 INFO:teuthology.orchestra.run.vm02.stdout:Get:60 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-cephfs amd64 20.2.0-712-g70f8415b-1jammy [184 kB] 2026-03-21T06:56:23.318 INFO:teuthology.orchestra.run.vm07.stdout:Get:64 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 20.2.0-712-g70f8415b-1jammy [112 kB] 2026-03-21T06:56:23.320 INFO:teuthology.orchestra.run.vm07.stdout:Get:65 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 20.2.0-712-g70f8415b-1jammy [261 kB] 2026-03-21T06:56:23.411 INFO:teuthology.orchestra.run.vm02.stdout:Get:61 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-ceph-common all 20.2.0-712-g70f8415b-1jammy [83.8 kB] 2026-03-21T06:56:23.412 INFO:teuthology.orchestra.run.vm02.stdout:Get:62 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rbd amd64 20.2.0-712-g70f8415b-1jammy [341 kB] 2026-03-21T06:56:23.437 INFO:teuthology.orchestra.run.vm07.stdout:Get:66 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 20.2.0-712-g70f8415b-1jammy [29.3 MB] 2026-03-21T06:56:23.533 INFO:teuthology.orchestra.run.vm02.stdout:Get:63 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 librgw2 amd64 20.2.0-712-g70f8415b-1jammy [8697 kB] 2026-03-21T06:56:23.580 INFO:teuthology.orchestra.run.vm04.stdout:Get:82 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 20.2.0-712-g70f8415b-1jammy [135 kB] 2026-03-21T06:56:23.580 INFO:teuthology.orchestra.run.vm04.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-daemon amd64 20.2.0-712-g70f8415b-1jammy [43.3 kB] 2026-03-21T06:56:23.581 INFO:teuthology.orchestra.run.vm04.stdout:Get:84 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-proxy2 amd64 20.2.0-712-g70f8415b-1jammy [30.7 kB] 2026-03-21T06:56:23.581 INFO:teuthology.orchestra.run.vm04.stdout:Get:85 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 20.2.0-712-g70f8415b-1jammy [41.5 kB] 2026-03-21T06:56:23.581 INFO:teuthology.orchestra.run.vm04.stdout:Get:86 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 20.2.0-712-g70f8415b-1jammy [25.1 MB] 2026-03-21T06:56:24.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:23 vm02 bash[17657]: cluster 2026-03-21T06:56:23.608916+0000 mgr.x (mgr.14152) 375 : cluster [DBG] pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:24.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:23 vm02 bash[17657]: cluster 2026-03-21T06:56:23.608916+0000 mgr.x (mgr.14152) 375 : cluster [DBG] pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:24.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:23 vm07 bash[19945]: cluster 2026-03-21T06:56:23.608916+0000 mgr.x (mgr.14152) 375 : cluster [DBG] pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:24.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:23 vm07 bash[19945]: cluster 2026-03-21T06:56:23.608916+0000 mgr.x (mgr.14152) 375 : cluster [DBG] pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:24.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:23 vm04 bash[20194]: cluster 2026-03-21T06:56:23.608916+0000 mgr.x (mgr.14152) 375 : cluster [DBG] pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:24.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:23 vm04 bash[20194]: cluster 2026-03-21T06:56:23.608916+0000 mgr.x (mgr.14152) 375 : cluster [DBG] pgmap v306: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:24.623 INFO:teuthology.orchestra.run.vm04.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 20.2.0-712-g70f8415b-1jammy [97.9 kB] 2026-03-21T06:56:24.866 INFO:teuthology.orchestra.run.vm04.stdout:Fetched 281 MB in 13s (21.9 MB/s) 2026-03-21T06:56:24.913 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-21T06:56:24.939 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 119262 files and directories currently installed.) 2026-03-21T06:56:24.941 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../00-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-21T06:56:24.943 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-21T06:56:24.962 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-21T06:56:24.967 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../01-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-21T06:56:24.968 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-21T06:56:24.982 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-21T06:56:24.987 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../02-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-21T06:56:24.988 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-21T06:56:25.008 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-21T06:56:25.013 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../03-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-21T06:56:25.017 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:25.059 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-21T06:56:25.065 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../04-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-21T06:56:25.067 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:25.084 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-21T06:56:25.090 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../05-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-21T06:56:25.091 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:25.114 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-21T06:56:25.119 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../06-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-21T06:56:25.120 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-21T06:56:25.144 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../07-librbd1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:25.146 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking librbd1 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-21T06:56:25.210 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../08-librados2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:25.213 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking librados2 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-21T06:56:25.276 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libnbd0. 2026-03-21T06:56:25.281 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../09-libnbd0_1.10.5-1_amd64.deb ... 2026-03-21T06:56:25.282 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-21T06:56:25.298 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libcephfs2. 2026-03-21T06:56:25.304 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../10-libcephfs2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:25.304 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:25.330 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-rados. 2026-03-21T06:56:25.336 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../11-python3-rados_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:25.336 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:25.355 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-21T06:56:25.361 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../12-python3-ceph-argparse_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:25.362 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:25.375 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-cephfs. 2026-03-21T06:56:25.381 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../13-python3-cephfs_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:25.382 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:25.400 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-21T06:56:25.406 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../14-python3-ceph-common_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:25.406 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:25.427 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-21T06:56:25.433 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../15-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-21T06:56:25.433 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-21T06:56:25.451 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-prettytable. 2026-03-21T06:56:25.456 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../16-python3-prettytable_2.5.0-2_all.deb ... 2026-03-21T06:56:25.457 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-21T06:56:25.471 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-rbd. 2026-03-21T06:56:25.477 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../17-python3-rbd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:25.478 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:25.500 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-21T06:56:25.506 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../18-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-21T06:56:25.507 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-21T06:56:25.531 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package librgw2. 2026-03-21T06:56:25.536 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../19-librgw2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:25.537 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:25.567 INFO:teuthology.orchestra.run.vm02.stdout:Get:64 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 python3-rgw amd64 20.2.0-712-g70f8415b-1jammy [112 kB] 2026-03-21T06:56:25.567 INFO:teuthology.orchestra.run.vm02.stdout:Get:65 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libradosstriper1 amd64 20.2.0-712-g70f8415b-1jammy [261 kB] 2026-03-21T06:56:25.568 INFO:teuthology.orchestra.run.vm02.stdout:Get:66 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-common amd64 20.2.0-712-g70f8415b-1jammy [29.3 MB] 2026-03-21T06:56:25.699 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-rgw. 2026-03-21T06:56:25.704 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../20-python3-rgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:25.705 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:25.731 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-21T06:56:25.737 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../21-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-21T06:56:25.738 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-21T06:56:25.754 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libradosstriper1. 2026-03-21T06:56:25.759 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../22-libradosstriper1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:25.760 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:25.780 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-common. 2026-03-21T06:56:25.786 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../23-ceph-common_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:25.786 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:26.221 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-base. 2026-03-21T06:56:26.227 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../24-ceph-base_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:26.232 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:26.340 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-21T06:56:26.347 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../25-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-21T06:56:26.348 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-21T06:56:26.362 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-cheroot. 2026-03-21T06:56:26.368 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../26-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-21T06:56:26.369 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-21T06:56:26.387 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-21T06:56:26.393 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../27-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-21T06:56:26.394 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-21T06:56:26.409 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-21T06:56:26.415 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../28-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-21T06:56:26.416 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-21T06:56:26.431 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-21T06:56:26.437 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../29-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-21T06:56:26.437 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-21T06:56:26.452 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-tempora. 2026-03-21T06:56:26.458 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../30-python3-tempora_4.1.2-1_all.deb ... 2026-03-21T06:56:26.458 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-21T06:56:26.473 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-portend. 2026-03-21T06:56:26.479 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../31-python3-portend_3.0.0-1_all.deb ... 2026-03-21T06:56:26.479 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-21T06:56:26.494 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-21T06:56:26.500 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../32-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-21T06:56:26.500 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-21T06:56:26.518 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-21T06:56:26.524 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../33-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-21T06:56:26.524 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-21T06:56:26.552 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-natsort. 2026-03-21T06:56:26.558 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../34-python3-natsort_8.0.2-1_all.deb ... 2026-03-21T06:56:26.559 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-21T06:56:26.575 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-21T06:56:26.581 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../35-ceph-mgr-modules-core_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:26.582 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:26.618 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-21T06:56:26.623 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../36-libsqlite3-mod-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:26.624 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:26.641 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-mgr. 2026-03-21T06:56:26.647 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../37-ceph-mgr_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:26.648 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:26.676 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-mon. 2026-03-21T06:56:26.681 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../38-ceph-mon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:26.682 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:26.776 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-21T06:56:26.782 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../39-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-21T06:56:26.783 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-21T06:56:26.800 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-osd. 2026-03-21T06:56:26.805 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../40-ceph-osd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:26.806 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:27.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:26 vm04 bash[20194]: cluster 2026-03-21T06:56:25.609263+0000 mgr.x (mgr.14152) 376 : cluster [DBG] pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:27.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:26 vm04 bash[20194]: cluster 2026-03-21T06:56:25.609263+0000 mgr.x (mgr.14152) 376 : cluster [DBG] pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:27.055 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph. 2026-03-21T06:56:27.062 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../41-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:27.063 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:27.078 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-fuse. 2026-03-21T06:56:27.084 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../42-ceph-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:27.084 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:27.112 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-mds. 2026-03-21T06:56:27.118 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../43-ceph-mds_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:27.119 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:27.139 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:26 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:56:27.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:26 vm02 bash[17657]: cluster 2026-03-21T06:56:25.609263+0000 mgr.x (mgr.14152) 376 : cluster [DBG] pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:27.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:26 vm02 bash[17657]: cluster 2026-03-21T06:56:25.609263+0000 mgr.x (mgr.14152) 376 : cluster [DBG] pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:26 vm07 bash[19945]: cluster 2026-03-21T06:56:25.609263+0000 mgr.x (mgr.14152) 376 : cluster [DBG] pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:26 vm07 bash[19945]: cluster 2026-03-21T06:56:25.609263+0000 mgr.x (mgr.14152) 376 : cluster [DBG] pgmap v307: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:27.163 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package cephadm. 2026-03-21T06:56:27.169 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../44-cephadm_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:27.170 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:27.189 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-21T06:56:27.195 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../45-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-21T06:56:27.196 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:27.221 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-21T06:56:27.226 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../46-ceph-mgr-cephadm_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:27.227 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:27.252 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-21T06:56:27.257 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../47-python3-repoze.lru_0.7-2_all.deb ... 2026-03-21T06:56:27.258 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-21T06:56:27.273 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-routes. 2026-03-21T06:56:27.278 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../48-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-21T06:56:27.279 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-21T06:56:27.301 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-21T06:56:27.306 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../49-ceph-mgr-dashboard_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:27.307 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:27.834 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:27 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:56:27.939 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-21T06:56:27.945 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../50-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-21T06:56:27.946 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-21T06:56:28.003 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-joblib. 2026-03-21T06:56:28.008 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../51-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-21T06:56:28.009 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-21T06:56:28.041 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-21T06:56:28.045 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../52-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-21T06:56:28.046 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-21T06:56:28.062 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-sklearn. 2026-03-21T06:56:28.066 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../53-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-21T06:56:28.067 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-21T06:56:28.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:27 vm02 bash[17657]: audit 2026-03-21T06:56:26.644553+0000 mgr.x (mgr.14152) 377 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:28.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:27 vm02 bash[17657]: audit 2026-03-21T06:56:26.644553+0000 mgr.x (mgr.14152) 377 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:28.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:27 vm07 bash[19945]: audit 2026-03-21T06:56:26.644553+0000 mgr.x (mgr.14152) 377 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:28.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:27 vm07 bash[19945]: audit 2026-03-21T06:56:26.644553+0000 mgr.x (mgr.14152) 377 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:28.192 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-21T06:56:28.199 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../54-ceph-mgr-diskprediction-local_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:28.200 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:28.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:27 vm04 bash[20194]: audit 2026-03-21T06:56:26.644553+0000 mgr.x (mgr.14152) 377 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:28.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:27 vm04 bash[20194]: audit 2026-03-21T06:56:26.644553+0000 mgr.x (mgr.14152) 377 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:28.468 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-cachetools. 2026-03-21T06:56:28.474 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../55-python3-cachetools_5.0.0-1_all.deb ... 2026-03-21T06:56:28.475 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-21T06:56:28.492 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-rsa. 2026-03-21T06:56:28.498 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../56-python3-rsa_4.8-1_all.deb ... 2026-03-21T06:56:28.499 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-21T06:56:28.518 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-google-auth. 2026-03-21T06:56:28.525 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../57-python3-google-auth_1.5.1-3_all.deb ... 2026-03-21T06:56:28.526 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-21T06:56:28.545 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-21T06:56:28.551 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../58-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-21T06:56:28.552 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-21T06:56:28.568 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-websocket. 2026-03-21T06:56:28.574 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../59-python3-websocket_1.2.3-1_all.deb ... 2026-03-21T06:56:28.574 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-21T06:56:28.593 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-21T06:56:28.598 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../60-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-21T06:56:28.599 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-21T06:56:28.738 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-21T06:56:28.745 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../61-ceph-mgr-k8sevents_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:28.746 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:28.763 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-21T06:56:28.768 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../62-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-21T06:56:28.769 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-21T06:56:28.789 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-21T06:56:28.795 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../63-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-21T06:56:28.796 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:28.804 INFO:teuthology.orchestra.run.vm02.stdout:Get:67 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 20.2.0-712-g70f8415b-1jammy [5415 kB] 2026-03-21T06:56:28.813 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package jq. 2026-03-21T06:56:28.819 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../64-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-21T06:56:28.820 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:28.872 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package socat. 2026-03-21T06:56:28.880 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../65-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-21T06:56:28.881 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-21T06:56:28.905 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package xmlstarlet. 2026-03-21T06:56:28.912 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../66-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-21T06:56:28.913 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-21T06:56:28.959 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-test. 2026-03-21T06:56:28.966 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../67-ceph-test_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:28.967 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:29.053 INFO:teuthology.orchestra.run.vm07.stdout:Get:67 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-base amd64 20.2.0-712-g70f8415b-1jammy [5415 kB] 2026-03-21T06:56:29.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:28 vm02 bash[17657]: audit 2026-03-21T06:56:27.480029+0000 mgr.x (mgr.14152) 378 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:29.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:28 vm02 bash[17657]: audit 2026-03-21T06:56:27.480029+0000 mgr.x (mgr.14152) 378 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:29.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:28 vm02 bash[17657]: cluster 2026-03-21T06:56:27.609641+0000 mgr.x (mgr.14152) 379 : cluster [DBG] pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:29.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:28 vm02 bash[17657]: cluster 2026-03-21T06:56:27.609641+0000 mgr.x (mgr.14152) 379 : cluster [DBG] pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:28 vm07 bash[19945]: audit 2026-03-21T06:56:27.480029+0000 mgr.x (mgr.14152) 378 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:28 vm07 bash[19945]: audit 2026-03-21T06:56:27.480029+0000 mgr.x (mgr.14152) 378 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:28 vm07 bash[19945]: cluster 2026-03-21T06:56:27.609641+0000 mgr.x (mgr.14152) 379 : cluster [DBG] pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:28 vm07 bash[19945]: cluster 2026-03-21T06:56:27.609641+0000 mgr.x (mgr.14152) 379 : cluster [DBG] pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:29.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:28 vm04 bash[20194]: audit 2026-03-21T06:56:27.480029+0000 mgr.x (mgr.14152) 378 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:29.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:28 vm04 bash[20194]: audit 2026-03-21T06:56:27.480029+0000 mgr.x (mgr.14152) 378 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:29.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:28 vm04 bash[20194]: cluster 2026-03-21T06:56:27.609641+0000 mgr.x (mgr.14152) 379 : cluster [DBG] pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:29.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:28 vm04 bash[20194]: cluster 2026-03-21T06:56:27.609641+0000 mgr.x (mgr.14152) 379 : cluster [DBG] pgmap v308: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:29.271 INFO:teuthology.orchestra.run.vm02.stdout:Get:68 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 20.2.0-712-g70f8415b-1jammy [246 kB] 2026-03-21T06:56:29.272 INFO:teuthology.orchestra.run.vm02.stdout:Get:69 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 20.2.0-712-g70f8415b-1jammy [124 kB] 2026-03-21T06:56:29.273 INFO:teuthology.orchestra.run.vm02.stdout:Get:70 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 20.2.0-712-g70f8415b-1jammy [906 kB] 2026-03-21T06:56:29.287 INFO:teuthology.orchestra.run.vm02.stdout:Get:71 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 20.2.0-712-g70f8415b-1jammy [6399 kB] 2026-03-21T06:56:29.638 INFO:teuthology.orchestra.run.vm07.stdout:Get:68 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-modules-core all 20.2.0-712-g70f8415b-1jammy [246 kB] 2026-03-21T06:56:29.646 INFO:teuthology.orchestra.run.vm07.stdout:Get:69 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libsqlite3-mod-ceph amd64 20.2.0-712-g70f8415b-1jammy [124 kB] 2026-03-21T06:56:29.649 INFO:teuthology.orchestra.run.vm07.stdout:Get:70 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr amd64 20.2.0-712-g70f8415b-1jammy [906 kB] 2026-03-21T06:56:29.755 INFO:teuthology.orchestra.run.vm02.stdout:Get:72 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 20.2.0-712-g70f8415b-1jammy [21.7 MB] 2026-03-21T06:56:29.758 INFO:teuthology.orchestra.run.vm07.stdout:Get:71 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mon amd64 20.2.0-712-g70f8415b-1jammy [6399 kB] 2026-03-21T06:56:30.333 INFO:teuthology.orchestra.run.vm07.stdout:Get:72 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-osd amd64 20.2.0-712-g70f8415b-1jammy [21.7 MB] 2026-03-21T06:56:30.424 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package ceph-volume. 2026-03-21T06:56:30.430 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../68-ceph-volume_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:30.431 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:30.458 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libcephfs-daemon. 2026-03-21T06:56:30.464 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../69-libcephfs-daemon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:30.465 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:30.481 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libcephfs-proxy2. 2026-03-21T06:56:30.487 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../70-libcephfs-proxy2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:30.488 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:30.503 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-21T06:56:30.509 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../71-libcephfs-dev_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:30.510 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:30.528 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package nvme-cli. 2026-03-21T06:56:30.534 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../72-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-21T06:56:30.535 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-21T06:56:30.575 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-21T06:56:30.582 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../73-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-21T06:56:30.583 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:30.626 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-21T06:56:30.632 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../74-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-21T06:56:30.633 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-21T06:56:30.649 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-pluggy. 2026-03-21T06:56:30.655 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../75-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-21T06:56:30.656 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-21T06:56:30.672 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-psutil. 2026-03-21T06:56:30.677 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../76-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-21T06:56:30.678 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-21T06:56:30.702 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-py. 2026-03-21T06:56:30.709 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../77-python3-py_1.10.0-1_all.deb ... 2026-03-21T06:56:30.710 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-21T06:56:30.735 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-pygments. 2026-03-21T06:56:30.741 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../78-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-21T06:56:30.742 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-21T06:56:30.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:30 vm04 bash[20194]: cluster 2026-03-21T06:56:29.610001+0000 mgr.x (mgr.14152) 380 : cluster [DBG] pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:30.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:30 vm04 bash[20194]: cluster 2026-03-21T06:56:29.610001+0000 mgr.x (mgr.14152) 380 : cluster [DBG] pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:30.802 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-toml. 2026-03-21T06:56:30.807 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../79-python3-toml_0.10.2-1_all.deb ... 2026-03-21T06:56:30.808 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-21T06:56:30.825 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-pytest. 2026-03-21T06:56:30.832 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../80-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-21T06:56:30.839 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-21T06:56:30.880 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-simplejson. 2026-03-21T06:56:30.883 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../81-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-21T06:56:30.884 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-21T06:56:30.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:30 vm02 bash[17657]: cluster 2026-03-21T06:56:29.610001+0000 mgr.x (mgr.14152) 380 : cluster [DBG] pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:30.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:30 vm02 bash[17657]: cluster 2026-03-21T06:56:29.610001+0000 mgr.x (mgr.14152) 380 : cluster [DBG] pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:30 vm07 bash[19945]: cluster 2026-03-21T06:56:29.610001+0000 mgr.x (mgr.14152) 380 : cluster [DBG] pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:30.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:30 vm07 bash[19945]: cluster 2026-03-21T06:56:29.610001+0000 mgr.x (mgr.14152) 380 : cluster [DBG] pgmap v309: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:30.905 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-webob. 2026-03-21T06:56:30.911 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../82-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-21T06:56:30.912 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-21T06:56:30.932 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-21T06:56:30.938 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../83-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-21T06:56:30.943 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-21T06:56:30.977 INFO:teuthology.orchestra.run.vm02.stdout:Get:73 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 20.2.0-712-g70f8415b-1jammy [14.1 kB] 2026-03-21T06:56:30.977 INFO:teuthology.orchestra.run.vm02.stdout:Get:74 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 20.2.0-712-g70f8415b-1jammy [955 kB] 2026-03-21T06:56:31.045 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package radosgw. 2026-03-21T06:56:31.052 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../84-radosgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:31.052 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:31.065 INFO:teuthology.orchestra.run.vm02.stdout:Get:75 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 20.2.0-712-g70f8415b-1jammy [2341 kB] 2026-03-21T06:56:31.184 INFO:teuthology.orchestra.run.vm02.stdout:Get:76 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 20.2.0-712-g70f8415b-1jammy [1049 kB] 2026-03-21T06:56:31.204 INFO:teuthology.orchestra.run.vm02.stdout:Get:77 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 20.2.0-712-g70f8415b-1jammy [179 kB] 2026-03-21T06:56:31.208 INFO:teuthology.orchestra.run.vm02.stdout:Get:78 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 20.2.0-712-g70f8415b-1jammy [45.5 MB] 2026-03-21T06:56:31.445 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package rbd-fuse. 2026-03-21T06:56:31.451 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../85-rbd-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:31.452 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:31.473 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package smartmontools. 2026-03-21T06:56:31.479 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../86-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-21T06:56:31.489 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-21T06:56:31.529 INFO:teuthology.orchestra.run.vm04.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-21T06:56:31.793 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-21T06:56:31.793 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-21T06:56:31.868 INFO:teuthology.orchestra.run.vm07.stdout:Get:73 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph amd64 20.2.0-712-g70f8415b-1jammy [14.1 kB] 2026-03-21T06:56:31.868 INFO:teuthology.orchestra.run.vm07.stdout:Get:74 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-fuse amd64 20.2.0-712-g70f8415b-1jammy [955 kB] 2026-03-21T06:56:31.909 INFO:teuthology.orchestra.run.vm07.stdout:Get:75 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mds amd64 20.2.0-712-g70f8415b-1jammy [2341 kB] 2026-03-21T06:56:31.939 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:31 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:31.939 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:31 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:31.939 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:31 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:31.939 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:31 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.058 INFO:teuthology.orchestra.run.vm07.stdout:Get:76 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 cephadm amd64 20.2.0-712-g70f8415b-1jammy [1049 kB] 2026-03-21T06:56:32.119 INFO:teuthology.orchestra.run.vm07.stdout:Get:77 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-cephadm all 20.2.0-712-g70f8415b-1jammy [179 kB] 2026-03-21T06:56:32.127 INFO:teuthology.orchestra.run.vm07.stdout:Get:78 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-dashboard all 20.2.0-712-g70f8415b-1jammy [45.5 MB] 2026-03-21T06:56:32.226 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:31 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.226 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.226 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:31 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.227 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.227 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:31 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.227 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.227 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:31 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.227 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.234 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-21T06:56:32.327 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-21T06:56:32.345 INFO:teuthology.orchestra.run.vm04.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-21T06:56:32.424 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-21T06:56:32.682 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-21T06:56:32.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:32 vm04 bash[20194]: cluster 2026-03-21T06:56:31.610330+0000 mgr.x (mgr.14152) 381 : cluster [DBG] pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:32.799 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:32 vm04 bash[20194]: cluster 2026-03-21T06:56:31.610330+0000 mgr.x (mgr.14152) 381 : cluster [DBG] pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:32.799 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.799 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:32.799 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:33.101 INFO:teuthology.orchestra.run.vm04.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-21T06:56:33.103 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:33.103 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:33 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:33.103 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:33.103 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:33 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:33.103 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:33.104 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:33 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:33.104 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:32 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:33.104 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:33 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:33.124 INFO:teuthology.orchestra.run.vm04.stdout:Setting up cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:33.136 INFO:teuthology.orchestra.run.vm02.stdout:Get:79 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 20.2.0-712-g70f8415b-1jammy [8625 kB] 2026-03-21T06:56:33.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:32 vm02 bash[17657]: cluster 2026-03-21T06:56:31.610330+0000 mgr.x (mgr.14152) 381 : cluster [DBG] pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:33.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:32 vm02 bash[17657]: cluster 2026-03-21T06:56:31.610330+0000 mgr.x (mgr.14152) 381 : cluster [DBG] pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:32 vm07 bash[19945]: cluster 2026-03-21T06:56:31.610330+0000 mgr.x (mgr.14152) 381 : cluster [DBG] pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:32 vm07 bash[19945]: cluster 2026-03-21T06:56:31.610330+0000 mgr.x (mgr.14152) 381 : cluster [DBG] pgmap v310: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:33.207 INFO:teuthology.orchestra.run.vm04.stdout:Adding system user cephadm....done 2026-03-21T06:56:33.232 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-21T06:56:33.311 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:33.318 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-21T06:56:33.390 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-21T06:56:33.492 INFO:teuthology.orchestra.run.vm02.stdout:Get:80 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 20.2.0-712-g70f8415b-1jammy [14.2 kB] 2026-03-21T06:56:33.492 INFO:teuthology.orchestra.run.vm02.stdout:Get:81 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 20.2.0-712-g70f8415b-1jammy [99.5 MB] 2026-03-21T06:56:33.497 INFO:teuthology.orchestra.run.vm04.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-21T06:56:33.510 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-21T06:56:33.599 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-21T06:56:33.743 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-21T06:56:33.842 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-21T06:56:33.925 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:33.993 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-21T06:56:34.002 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-21T06:56:34.007 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-21T06:56:34.014 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-21T06:56:34.037 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-21T06:56:34.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:33 vm02 bash[17657]: cluster 2026-03-21T06:56:33.610639+0000 mgr.x (mgr.14152) 382 : cluster [DBG] pgmap v311: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:34.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:33 vm02 bash[17657]: cluster 2026-03-21T06:56:33.610639+0000 mgr.x (mgr.14152) 382 : cluster [DBG] pgmap v311: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:34.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:33 vm07 bash[19945]: cluster 2026-03-21T06:56:33.610639+0000 mgr.x (mgr.14152) 382 : cluster [DBG] pgmap v311: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:34.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:33 vm07 bash[19945]: cluster 2026-03-21T06:56:33.610639+0000 mgr.x (mgr.14152) 382 : cluster [DBG] pgmap v311: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:34.176 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-21T06:56:34.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:33 vm04 bash[20194]: cluster 2026-03-21T06:56:33.610639+0000 mgr.x (mgr.14152) 382 : cluster [DBG] pgmap v311: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:34.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:33 vm04 bash[20194]: cluster 2026-03-21T06:56:33.610639+0000 mgr.x (mgr.14152) 382 : cluster [DBG] pgmap v311: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:34.274 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:34.285 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-21T06:56:34.360 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-21T06:56:34.439 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-21T06:56:34.529 INFO:teuthology.orchestra.run.vm07.stdout:Get:79 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-diskprediction-local all 20.2.0-712-g70f8415b-1jammy [8625 kB] 2026-03-21T06:56:34.712 INFO:teuthology.orchestra.run.vm04.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-21T06:56:34.719 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-21T06:56:34.821 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:34.984 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-21T06:56:34.999 INFO:teuthology.orchestra.run.vm07.stdout:Get:80 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-mgr-k8sevents all 20.2.0-712-g70f8415b-1jammy [14.2 kB] 2026-03-21T06:56:34.999 INFO:teuthology.orchestra.run.vm07.stdout:Get:81 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-test amd64 20.2.0-712-g70f8415b-1jammy [99.5 MB] 2026-03-21T06:56:35.078 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-21T06:56:35.166 INFO:teuthology.orchestra.run.vm04.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-21T06:56:35.186 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:35.288 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-21T06:56:36.084 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:36.102 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-21T06:56:36.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:35 vm02 bash[17657]: cluster 2026-03-21T06:56:35.611013+0000 mgr.x (mgr.14152) 383 : cluster [DBG] pgmap v312: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:36.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:35 vm02 bash[17657]: cluster 2026-03-21T06:56:35.611013+0000 mgr.x (mgr.14152) 383 : cluster [DBG] pgmap v312: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:36.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:35 vm07 bash[19945]: cluster 2026-03-21T06:56:35.611013+0000 mgr.x (mgr.14152) 383 : cluster [DBG] pgmap v312: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:36.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:35 vm07 bash[19945]: cluster 2026-03-21T06:56:35.611013+0000 mgr.x (mgr.14152) 383 : cluster [DBG] pgmap v312: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:36.181 INFO:teuthology.orchestra.run.vm04.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-21T06:56:36.193 INFO:teuthology.orchestra.run.vm04.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-21T06:56:36.207 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-21T06:56:36.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:35 vm04 bash[20194]: cluster 2026-03-21T06:56:35.611013+0000 mgr.x (mgr.14152) 383 : cluster [DBG] pgmap v312: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:36.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:35 vm04 bash[20194]: cluster 2026-03-21T06:56:35.611013+0000 mgr.x (mgr.14152) 383 : cluster [DBG] pgmap v312: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:36.295 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-21T06:56:36.381 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:36.390 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-21T06:56:36.473 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-21T06:56:36.552 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-21T06:56:36.637 INFO:teuthology.orchestra.run.vm04.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-21T06:56:36.644 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-21T06:56:36.730 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-21T06:56:36.746 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-21T06:56:36.818 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-21T06:56:36.916 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-21T06:56:36.997 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:37.011 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-21T06:56:37.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:36 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:56:37.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:36 vm02 bash[17657]: audit 2026-03-21T06:56:36.654831+0000 mgr.x (mgr.14152) 384 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:37.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:36 vm02 bash[17657]: audit 2026-03-21T06:56:36.654831+0000 mgr.x (mgr.14152) 384 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:36 vm07 bash[19945]: audit 2026-03-21T06:56:36.654831+0000 mgr.x (mgr.14152) 384 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:36 vm07 bash[19945]: audit 2026-03-21T06:56:36.654831+0000 mgr.x (mgr.14152) 384 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:37.154 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-21T06:56:37.246 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:37.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:36 vm04 bash[20194]: audit 2026-03-21T06:56:36.654831+0000 mgr.x (mgr.14152) 384 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:37.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:36 vm04 bash[20194]: audit 2026-03-21T06:56:36.654831+0000 mgr.x (mgr.14152) 384 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:37.262 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-21T06:56:37.365 INFO:teuthology.orchestra.run.vm04.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:37.373 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-21T06:56:37.511 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-21T06:56:37.524 INFO:teuthology.orchestra.run.vm04.stdout:Setting up librados2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:37.530 INFO:teuthology.orchestra.run.vm04.stdout:Setting up librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:37.533 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:37.555 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-21T06:56:37.879 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:37 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:56:37.879 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:37 vm07 bash[19945]: audit 2026-03-21T06:56:37.491127+0000 mgr.x (mgr.14152) 385 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:37.879 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:37 vm07 bash[19945]: audit 2026-03-21T06:56:37.491127+0000 mgr.x (mgr.14152) 385 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:37.879 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:37 vm07 bash[19945]: cluster 2026-03-21T06:56:37.611394+0000 mgr.x (mgr.14152) 386 : cluster [DBG] pgmap v313: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:37.879 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:37 vm07 bash[19945]: cluster 2026-03-21T06:56:37.611394+0000 mgr.x (mgr.14152) 386 : cluster [DBG] pgmap v313: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:38.042 INFO:teuthology.orchestra.run.vm02.stdout:Get:82 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 20.2.0-712-g70f8415b-1jammy [135 kB] 2026-03-21T06:56:38.042 INFO:teuthology.orchestra.run.vm02.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-daemon amd64 20.2.0-712-g70f8415b-1jammy [43.3 kB] 2026-03-21T06:56:38.042 INFO:teuthology.orchestra.run.vm02.stdout:Get:84 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-proxy2 amd64 20.2.0-712-g70f8415b-1jammy [30.7 kB] 2026-03-21T06:56:38.042 INFO:teuthology.orchestra.run.vm02.stdout:Get:85 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 20.2.0-712-g70f8415b-1jammy [41.5 kB] 2026-03-21T06:56:38.043 INFO:teuthology.orchestra.run.vm02.stdout:Get:86 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 20.2.0-712-g70f8415b-1jammy [25.1 MB] 2026-03-21T06:56:38.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:37 vm02 bash[17657]: audit 2026-03-21T06:56:37.491127+0000 mgr.x (mgr.14152) 385 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:38.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:37 vm02 bash[17657]: audit 2026-03-21T06:56:37.491127+0000 mgr.x (mgr.14152) 385 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:38.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:37 vm02 bash[17657]: cluster 2026-03-21T06:56:37.611394+0000 mgr.x (mgr.14152) 386 : cluster [DBG] pgmap v313: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:38.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:37 vm02 bash[17657]: cluster 2026-03-21T06:56:37.611394+0000 mgr.x (mgr.14152) 386 : cluster [DBG] pgmap v313: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:38.152 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.165 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.177 INFO:teuthology.orchestra.run.vm04.stdout:Setting up librbd1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.191 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.203 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:37 vm04 bash[20194]: audit 2026-03-21T06:56:37.491127+0000 mgr.x (mgr.14152) 385 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:38.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:37 vm04 bash[20194]: audit 2026-03-21T06:56:37.491127+0000 mgr.x (mgr.14152) 385 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:38.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:37 vm04 bash[20194]: cluster 2026-03-21T06:56:37.611394+0000 mgr.x (mgr.14152) 386 : cluster [DBG] pgmap v313: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:38.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:37 vm04 bash[20194]: cluster 2026-03-21T06:56:37.611394+0000 mgr.x (mgr.14152) 386 : cluster [DBG] pgmap v313: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:38.272 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-21T06:56:38.272 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-21T06:56:38.602 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:38.602 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:38.602 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:38.602 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:38.715 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.728 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.742 INFO:teuthology.orchestra.run.vm04.stdout:Setting up libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.750 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.768 INFO:teuthology.orchestra.run.vm04.stdout:Setting up rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.781 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.793 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.807 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:38.872 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:38.872 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:38.873 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:38.873 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:38.877 INFO:teuthology.orchestra.run.vm04.stdout:Adding group ceph....done 2026-03-21T06:56:38.950 INFO:teuthology.orchestra.run.vm04.stdout:Adding system user ceph....done 2026-03-21T06:56:38.985 INFO:teuthology.orchestra.run.vm04.stdout:Setting system user ceph properties....done 2026-03-21T06:56:38.991 INFO:teuthology.orchestra.run.vm04.stdout:Fixing /var/run/ceph ownership....done 2026-03-21T06:56:39.096 INFO:teuthology.orchestra.run.vm02.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 20.2.0-712-g70f8415b-1jammy [97.9 kB] 2026-03-21T06:56:39.345 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-21T06:56:39.381 INFO:teuthology.orchestra.run.vm02.stdout:Fetched 281 MB in 27s (10.2 MB/s) 2026-03-21T06:56:39.456 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.456 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.456 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.456 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.572 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-21T06:56:39.602 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 119262 files and directories currently installed.) 2026-03-21T06:56:39.604 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../00-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-21T06:56:39.607 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-21T06:56:39.643 INFO:teuthology.orchestra.run.vm07.stdout:Get:82 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 ceph-volume all 20.2.0-712-g70f8415b-1jammy [135 kB] 2026-03-21T06:56:39.644 INFO:teuthology.orchestra.run.vm07.stdout:Get:83 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-daemon amd64 20.2.0-712-g70f8415b-1jammy [43.3 kB] 2026-03-21T06:56:39.644 INFO:teuthology.orchestra.run.vm07.stdout:Get:84 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-proxy2 amd64 20.2.0-712-g70f8415b-1jammy [30.7 kB] 2026-03-21T06:56:39.644 INFO:teuthology.orchestra.run.vm07.stdout:Get:85 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 libcephfs-dev amd64 20.2.0-712-g70f8415b-1jammy [41.5 kB] 2026-03-21T06:56:39.735 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.736 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.736 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.736 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.736 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.736 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.736 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.736 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:39.746 INFO:teuthology.orchestra.run.vm07.stdout:Get:86 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 radosgw amd64 20.2.0-712-g70f8415b-1jammy [25.1 MB] 2026-03-21T06:56:39.761 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:39.776 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-21T06:56:39.783 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../01-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-21T06:56:39.788 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-21T06:56:39.799 INFO:teuthology.orchestra.run.vm04.stdout:Setting up radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:39.844 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-21T06:56:39.850 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../02-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-21T06:56:39.851 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-21T06:56:39.900 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-21T06:56:39.907 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../03-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-21T06:56:39.912 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:39.978 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-21T06:56:39.985 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../04-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-21T06:56:39.986 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:39 vm04 bash[20194]: cluster 2026-03-21T06:56:39.611757+0000 mgr.x (mgr.14152) 387 : cluster [DBG] pgmap v314: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:39 vm04 bash[20194]: cluster 2026-03-21T06:56:39.611757+0000 mgr.x (mgr.14152) 387 : cluster [DBG] pgmap v314: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:40.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.009 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.009 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.009 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.037 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-21T06:56:40.043 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../05-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-21T06:56:40.048 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:40.078 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-21T06:56:40.078 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-21T06:56:40.111 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-21T06:56:40.116 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../06-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-21T06:56:40.118 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-21T06:56:40.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:39 vm02 bash[17657]: cluster 2026-03-21T06:56:39.611757+0000 mgr.x (mgr.14152) 387 : cluster [DBG] pgmap v314: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:40.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:39 vm02 bash[17657]: cluster 2026-03-21T06:56:39.611757+0000 mgr.x (mgr.14152) 387 : cluster [DBG] pgmap v314: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:40.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:39 vm07 bash[19945]: cluster 2026-03-21T06:56:39.611757+0000 mgr.x (mgr.14152) 387 : cluster [DBG] pgmap v314: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:40.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:39 vm07 bash[19945]: cluster 2026-03-21T06:56:39.611757+0000 mgr.x (mgr.14152) 387 : cluster [DBG] pgmap v314: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:40.172 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../07-librbd1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:40.179 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librbd1 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-21T06:56:40.275 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../08-librados2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:40.293 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librados2 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-21T06:56:40.379 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.379 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.379 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.379 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.379 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libnbd0. 2026-03-21T06:56:40.386 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../09-libnbd0_1.10.5-1_amd64.deb ... 2026-03-21T06:56:40.390 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-21T06:56:40.435 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libcephfs2. 2026-03-21T06:56:40.440 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../10-libcephfs2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:40.448 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:40.480 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rados. 2026-03-21T06:56:40.481 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:40.487 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../11-python3-rados_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:40.488 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:40.535 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-21T06:56:40.539 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../12-python3-ceph-argparse_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:40.544 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:40.588 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cephfs. 2026-03-21T06:56:40.594 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../13-python3-cephfs_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:40.595 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:40.599 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-21T06:56:40.635 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-21T06:56:40.641 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../14-python3-ceph-common_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:40.644 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:40.703 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-21T06:56:40.703 INFO:teuthology.orchestra.run.vm07.stdout:Get:87 https://1.chacra.ceph.com/r/ceph/tentacle/70f8415b300f041766fa27faf7d5472699e32388/ubuntu/jammy/flavors/default jammy/main amd64 rbd-fuse amd64 20.2.0-712-g70f8415b-1jammy [97.9 kB] 2026-03-21T06:56:40.710 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../15-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-21T06:56:40.717 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.717 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.717 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.717 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:40.718 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-21T06:56:40.744 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-prettytable. 2026-03-21T06:56:40.750 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../16-python3-prettytable_2.5.0-2_all.deb ... 2026-03-21T06:56:40.786 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-21T06:56:40.826 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rbd. 2026-03-21T06:56:40.833 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../17-python3-rbd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:40.839 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:40.893 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-21T06:56:40.899 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../18-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-21T06:56:40.905 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-21T06:56:40.954 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package librgw2. 2026-03-21T06:56:40.960 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../19-librgw2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:40.973 INFO:teuthology.orchestra.run.vm07.stdout:Fetched 281 MB in 29s (9663 kB/s) 2026-03-21T06:56:40.975 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:41.008 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.008 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.008 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.008 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.008 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.008 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:40 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.034 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:41.054 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package liblttng-ust1:amd64. 2026-03-21T06:56:41.085 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 119262 files and directories currently installed.) 2026-03-21T06:56:41.087 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../00-liblttng-ust1_2.13.1-1ubuntu1_amd64.deb ... 2026-03-21T06:56:41.112 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-21T06:56:41.112 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-21T06:56:41.123 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-21T06:56:41.160 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rgw. 2026-03-21T06:56:41.167 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../20-python3-rgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:41.169 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:41.174 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libdouble-conversion3:amd64. 2026-03-21T06:56:41.178 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../01-libdouble-conversion3_3.1.7-4_amd64.deb ... 2026-03-21T06:56:41.195 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-21T06:56:41.211 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-21T06:56:41.218 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../21-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-21T06:56:41.219 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-21T06:56:41.222 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libpcre2-16-0:amd64. 2026-03-21T06:56:41.226 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../02-libpcre2-16-0_10.39-3ubuntu0.1_amd64.deb ... 2026-03-21T06:56:41.239 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-21T06:56:41.261 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libradosstriper1. 2026-03-21T06:56:41.267 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../22-libradosstriper1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:41.268 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:41.280 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.281 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.281 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.281 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.310 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libqt5core5a:amd64. 2026-03-21T06:56:41.315 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../03-libqt5core5a_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-21T06:56:41.316 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-common. 2026-03-21T06:56:41.319 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:41.322 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../23-ceph-common_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:41.324 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:41.452 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libqt5dbus5:amd64. 2026-03-21T06:56:41.458 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../04-libqt5dbus5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-21T06:56:41.460 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:41.515 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libqt5network5:amd64. 2026-03-21T06:56:41.521 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../05-libqt5network5_5.15.3+dfsg-2ubuntu0.2_amd64.deb ... 2026-03-21T06:56:41.523 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:41.553 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.553 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:41 vm04 bash[20194]: audit 2026-03-21T06:56:41.279241+0000 mon.b (mon.2) 22 : audit [DBG] from='client.? 192.168.123.104:0/1542676467' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:56:41.553 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:41 vm04 bash[20194]: audit 2026-03-21T06:56:41.279241+0000 mon.b (mon.2) 22 : audit [DBG] from='client.? 192.168.123.104:0/1542676467' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:56:41.553 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.553 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.553 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.560 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:41.594 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libthrift-0.16.0:amd64. 2026-03-21T06:56:41.601 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../06-libthrift-0.16.0_0.16.0-2_amd64.deb ... 2026-03-21T06:56:41.611 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-21T06:56:41.661 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-21T06:56:41.661 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-21T06:56:41.770 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../07-librbd1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:41.800 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking librbd1 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-21T06:56:41.803 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-base. 2026-03-21T06:56:41.809 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../24-ceph-base_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:41.815 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:41.842 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.843 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.843 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.843 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:41.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:41 vm02 bash[17657]: audit 2026-03-21T06:56:41.279241+0000 mon.b (mon.2) 22 : audit [DBG] from='client.? 192.168.123.104:0/1542676467' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:56:41.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:41 vm02 bash[17657]: audit 2026-03-21T06:56:41.279241+0000 mon.b (mon.2) 22 : audit [DBG] from='client.? 192.168.123.104:0/1542676467' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:56:41.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:41 vm07 bash[19945]: audit 2026-03-21T06:56:41.279241+0000 mon.b (mon.2) 22 : audit [DBG] from='client.? 192.168.123.104:0/1542676467' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:56:41.917 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:41 vm07 bash[19945]: audit 2026-03-21T06:56:41.279241+0000 mon.b (mon.2) 22 : audit [DBG] from='client.? 192.168.123.104:0/1542676467' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:56:41.946 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../08-librados2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:41.953 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-21T06:56:41.955 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking librados2 (20.2.0-712-g70f8415b-1jammy) over (17.2.9-0ubuntu0.22.04.2) ... 2026-03-21T06:56:41.959 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../25-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-21T06:56:41.960 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-21T06:56:41.998 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cheroot. 2026-03-21T06:56:42.004 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../26-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-21T06:56:42.026 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-21T06:56:42.059 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.062 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-21T06:56:42.062 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libnbd0. 2026-03-21T06:56:42.069 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../27-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-21T06:56:42.079 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../09-libnbd0_1.10.5-1_amd64.deb ... 2026-03-21T06:56:42.080 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-21T06:56:42.081 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libnbd0 (1.10.5-1) ... 2026-03-21T06:56:42.127 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-21T06:56:42.128 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libcephfs2. 2026-03-21T06:56:42.133 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../28-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-21T06:56:42.134 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-21T06:56:42.135 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../10-libcephfs2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:42.136 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.136 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-21T06:56:42.136 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-21T06:56:42.137 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.137 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.137 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.137 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.184 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-21T06:56:42.185 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-rados. 2026-03-21T06:56:42.190 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../29-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-21T06:56:42.195 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../11-python3-rados_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:42.195 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-21T06:56:42.196 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.237 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-tempora. 2026-03-21T06:56:42.243 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../30-python3-tempora_4.1.2-1_all.deb ... 2026-03-21T06:56:42.246 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-21T06:56:42.255 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-ceph-argparse. 2026-03-21T06:56:42.261 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../12-python3-ceph-argparse_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:42.265 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.273 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-portend. 2026-03-21T06:56:42.279 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../31-python3-portend_3.0.0-1_all.deb ... 2026-03-21T06:56:42.280 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-21T06:56:42.300 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-cephfs. 2026-03-21T06:56:42.306 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../13-python3-cephfs_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:42.320 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.330 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-21T06:56:42.336 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../32-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-21T06:56:42.341 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-21T06:56:42.445 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-ceph-common. 2026-03-21T06:56:42.446 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-21T06:56:42.453 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../33-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-21T06:56:42.454 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../14-python3-ceph-common_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:42.457 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-21T06:56:42.457 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.459 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.459 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.459 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.459 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.501 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-wcwidth. 2026-03-21T06:56:42.507 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../15-python3-wcwidth_0.2.5+dfsg1-1_all.deb ... 2026-03-21T06:56:42.508 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-21T06:56:42.521 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-natsort. 2026-03-21T06:56:42.527 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../34-python3-natsort_8.0.2-1_all.deb ... 2026-03-21T06:56:42.539 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-21T06:56:42.568 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.568 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-prettytable. 2026-03-21T06:56:42.574 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.574 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-21T06:56:42.574 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../16-python3-prettytable_2.5.0-2_all.deb ... 2026-03-21T06:56:42.580 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../35-ceph-mgr-modules-core_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:42.581 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-prettytable (2.5.0-2) ... 2026-03-21T06:56:42.581 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.602 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.633 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-rbd. 2026-03-21T06:56:42.639 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../17-python3-rbd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:42.640 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.642 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-21T06:56:42.648 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../36-libsqlite3-mod-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:42.650 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.680 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-21T06:56:42.680 INFO:teuthology.orchestra.run.vm04.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-21T06:56:42.680 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr. 2026-03-21T06:56:42.686 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../37-ceph-mgr_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:42.686 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package librdkafka1:amd64. 2026-03-21T06:56:42.686 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../18-librdkafka1_1.8.0-1build1_amd64.deb ... 2026-03-21T06:56:42.692 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.693 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-21T06:56:42.741 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package librgw2. 2026-03-21T06:56:42.748 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mon. 2026-03-21T06:56:42.749 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../19-librgw2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:42.752 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.755 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../38-ceph-mon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:42.758 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.758 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.758 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:42 vm04 bash[20194]: cluster 2026-03-21T06:56:41.612124+0000 mgr.x (mgr.14152) 388 : cluster [DBG] pgmap v315: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:42.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:42 vm04 bash[20194]: cluster 2026-03-21T06:56:41.612124+0000 mgr.x (mgr.14152) 388 : cluster [DBG] pgmap v315: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:42.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:42.765 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:42 vm02 bash[17657]: cluster 2026-03-21T06:56:41.612124+0000 mgr.x (mgr.14152) 388 : cluster [DBG] pgmap v315: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:42.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:42 vm02 bash[17657]: cluster 2026-03-21T06:56:41.612124+0000 mgr.x (mgr.14152) 388 : cluster [DBG] pgmap v315: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:42.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:42 vm07 bash[19945]: cluster 2026-03-21T06:56:41.612124+0000 mgr.x (mgr.14152) 388 : cluster [DBG] pgmap v315: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:42.905 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:42 vm07 bash[19945]: cluster 2026-03-21T06:56:41.612124+0000 mgr.x (mgr.14152) 388 : cluster [DBG] pgmap v315: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:42.941 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-21T06:56:42.942 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-rgw. 2026-03-21T06:56:42.948 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../39-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-21T06:56:42.948 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../20-python3-rgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:42.950 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-21T06:56:42.958 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.981 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-osd. 2026-03-21T06:56:42.987 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../40-ceph-osd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:42.988 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:42.990 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package liboath0:amd64. 2026-03-21T06:56:42.996 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../21-liboath0_2.6.7-3ubuntu0.1_amd64.deb ... 2026-03-21T06:56:43.012 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-21T06:56:43.058 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libradosstriper1. 2026-03-21T06:56:43.064 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../22-libradosstriper1_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:43.067 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.083 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.105 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:43.105 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:43.105 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:43.105 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:43.105 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:43.105 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:43.105 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:43.105 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:42 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:43.108 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.116 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.119 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-common. 2026-03-21T06:56:43.125 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../23-ceph-common_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:43.126 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.141 INFO:teuthology.orchestra.run.vm04.stdout:Setting up ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.273 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph. 2026-03-21T06:56:43.273 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T06:56:43.279 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../41-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:43.291 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.350 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-fuse. 2026-03-21T06:56:43.357 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../42-ceph-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:43.358 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.409 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mds. 2026-03-21T06:56:43.416 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../43-ceph-mds_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:43.420 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.438 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T06:56:43.479 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package cephadm. 2026-03-21T06:56:43.485 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../44-cephadm_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:43.488 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.522 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-21T06:56:43.529 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../45-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-21T06:56:43.538 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:43.721 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-21T06:56:43.727 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../46-ceph-mgr-cephadm_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:43.732 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.801 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-21T06:56:43.806 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../47-python3-repoze.lru_0.7-2_all.deb ... 2026-03-21T06:56:43.808 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-21T06:56:43.809 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-base. 2026-03-21T06:56:43.815 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../24-ceph-base_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:43.821 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.844 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-routes. 2026-03-21T06:56:43.851 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../48-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-21T06:56:43.856 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-21T06:56:43.923 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-21T06:56:43.929 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../49-ceph-mgr-dashboard_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:43.934 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:43.950 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:43.950 INFO:teuthology.orchestra.run.vm04.stdout:Running kernel seems to be up-to-date. 2026-03-21T06:56:43.950 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:43.950 INFO:teuthology.orchestra.run.vm04.stdout:Services to be restarted: 2026-03-21T06:56:43.954 INFO:teuthology.orchestra.run.vm04.stdout: systemctl restart apache-htcacheclean.service 2026-03-21T06:56:43.962 INFO:teuthology.orchestra.run.vm04.stdout: systemctl restart rsyslog.service 2026-03-21T06:56:43.965 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:43.966 INFO:teuthology.orchestra.run.vm04.stdout:Service restarts being deferred: 2026-03-21T06:56:43.966 INFO:teuthology.orchestra.run.vm04.stdout: systemctl restart networkd-dispatcher.service 2026-03-21T06:56:43.966 INFO:teuthology.orchestra.run.vm04.stdout: systemctl restart unattended-upgrades.service 2026-03-21T06:56:43.966 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:43.966 INFO:teuthology.orchestra.run.vm04.stdout:No containers need to be restarted. 2026-03-21T06:56:43.966 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:43.966 INFO:teuthology.orchestra.run.vm04.stdout:No user sessions are running outdated binaries. 2026-03-21T06:56:43.966 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:43.966 INFO:teuthology.orchestra.run.vm04.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-21T06:56:44.101 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-jaraco.functools. 2026-03-21T06:56:44.108 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../25-python3-jaraco.functools_3.4.0-2_all.deb ... 2026-03-21T06:56:44.109 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-jaraco.functools (3.4.0-2) ... 2026-03-21T06:56:44.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:43 vm02 bash[17657]: cluster 2026-03-21T06:56:43.612513+0000 mgr.x (mgr.14152) 389 : cluster [DBG] pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:44.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:43 vm02 bash[17657]: cluster 2026-03-21T06:56:43.612513+0000 mgr.x (mgr.14152) 389 : cluster [DBG] pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:44.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:43 vm07 bash[19945]: cluster 2026-03-21T06:56:43.612513+0000 mgr.x (mgr.14152) 389 : cluster [DBG] pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:44.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:43 vm07 bash[19945]: cluster 2026-03-21T06:56:43.612513+0000 mgr.x (mgr.14152) 389 : cluster [DBG] pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:44.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:43 vm04 bash[20194]: cluster 2026-03-21T06:56:43.612513+0000 mgr.x (mgr.14152) 389 : cluster [DBG] pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:44.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:43 vm04 bash[20194]: cluster 2026-03-21T06:56:43.612513+0000 mgr.x (mgr.14152) 389 : cluster [DBG] pgmap v316: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:44.576 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-cheroot. 2026-03-21T06:56:44.583 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../26-python3-cheroot_8.5.2+ds1-1ubuntu3.1_all.deb ... 2026-03-21T06:56:44.588 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-21T06:56:44.658 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-jaraco.classes. 2026-03-21T06:56:44.665 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../27-python3-jaraco.classes_3.2.1-3_all.deb ... 2026-03-21T06:56:44.668 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-jaraco.classes (3.2.1-3) ... 2026-03-21T06:56:44.706 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-jaraco.text. 2026-03-21T06:56:44.712 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../28-python3-jaraco.text_3.6.0-2_all.deb ... 2026-03-21T06:56:44.713 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-jaraco.text (3.6.0-2) ... 2026-03-21T06:56:44.841 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T06:56:44.843 DEBUG:teuthology.orchestra.run.vm04:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install open-iscsi multipath-tools python3-jmespath python3-xmltodict s3cmd 2026-03-21T06:56:44.907 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-jaraco.collections. 2026-03-21T06:56:44.913 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../29-python3-jaraco.collections_3.4.0-2_all.deb ... 2026-03-21T06:56:44.914 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-jaraco.collections (3.4.0-2) ... 2026-03-21T06:56:44.921 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T06:56:44.952 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-tempora. 2026-03-21T06:56:44.958 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../30-python3-tempora_4.1.2-1_all.deb ... 2026-03-21T06:56:44.964 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-tempora (4.1.2-1) ... 2026-03-21T06:56:45.001 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-21T06:56:45.005 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-portend. 2026-03-21T06:56:45.007 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../50-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-21T06:56:45.011 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../31-python3-portend_3.0.0-1_all.deb ... 2026-03-21T06:56:45.015 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-portend (3.0.0-1) ... 2026-03-21T06:56:45.015 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-21T06:56:45.075 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-zc.lockfile. 2026-03-21T06:56:45.081 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../32-python3-zc.lockfile_2.0-1_all.deb ... 2026-03-21T06:56:45.082 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-zc.lockfile (2.0-1) ... 2026-03-21T06:56:45.105 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T06:56:45.106 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T06:56:45.106 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-joblib. 2026-03-21T06:56:45.114 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../51-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-21T06:56:45.120 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-21T06:56:45.120 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-cherrypy3. 2026-03-21T06:56:45.128 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../33-python3-cherrypy3_18.6.1-4_all.deb ... 2026-03-21T06:56:45.132 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-cherrypy3 (18.6.1-4) ... 2026-03-21T06:56:45.185 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-natsort. 2026-03-21T06:56:45.188 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-21T06:56:45.191 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../52-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-21T06:56:45.195 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../34-python3-natsort_8.0.2-1_all.deb ... 2026-03-21T06:56:45.201 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-21T06:56:45.202 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-natsort (8.0.2-1) ... 2026-03-21T06:56:45.243 INFO:teuthology.orchestra.run.vm04.stdout:open-iscsi is already the newest version (2.1.5-1ubuntu1.1). 2026-03-21T06:56:45.243 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T06:56:45.243 INFO:teuthology.orchestra.run.vm04.stdout: libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T06:56:45.243 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T06:56:45.244 INFO:teuthology.orchestra.run.vm04.stdout:Suggested packages: 2026-03-21T06:56:45.244 INFO:teuthology.orchestra.run.vm04.stdout: multipath-tools-boot 2026-03-21T06:56:45.252 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-mgr-modules-core. 2026-03-21T06:56:45.256 INFO:teuthology.orchestra.run.vm04.stdout:The following NEW packages will be installed: 2026-03-21T06:56:45.256 INFO:teuthology.orchestra.run.vm04.stdout: multipath-tools python3-jmespath python3-xmltodict s3cmd 2026-03-21T06:56:45.258 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../35-ceph-mgr-modules-core_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:45.259 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:45.259 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-sklearn. 2026-03-21T06:56:45.267 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../53-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-21T06:56:45.278 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-21T06:56:45.330 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libsqlite3-mod-ceph. 2026-03-21T06:56:45.334 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../36-libsqlite3-mod-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:45.336 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:45.397 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-mgr. 2026-03-21T06:56:45.402 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../37-ceph-mgr_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:45.403 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:45.417 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-21T06:56:45.423 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../54-ceph-mgr-diskprediction-local_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:45.437 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:45.461 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 4 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T06:56:45.461 INFO:teuthology.orchestra.run.vm04.stdout:Need to get 486 kB of archives. 2026-03-21T06:56:45.461 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 1930 kB of additional disk space will be used. 2026-03-21T06:56:45.461 INFO:teuthology.orchestra.run.vm04.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-21T06:56:45.468 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-mon. 2026-03-21T06:56:45.474 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../38-ceph-mon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:45.478 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:45.666 INFO:teuthology.orchestra.run.vm04.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-21T06:56:45.689 INFO:teuthology.orchestra.run.vm04.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy/universe amd64 s3cmd all 2.2.0-1 [120 kB] 2026-03-21T06:56:45.730 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-cachetools. 2026-03-21T06:56:45.730 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libfuse2:amd64. 2026-03-21T06:56:45.736 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../39-libfuse2_2.9.9-5ubuntu3_amd64.deb ... 2026-03-21T06:56:45.737 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../55-python3-cachetools_5.0.0-1_all.deb ... 2026-03-21T06:56:45.737 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-21T06:56:45.738 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-21T06:56:45.781 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-rsa. 2026-03-21T06:56:45.783 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-osd. 2026-03-21T06:56:45.787 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../56-python3-rsa_4.8-1_all.deb ... 2026-03-21T06:56:45.790 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-21T06:56:45.790 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../40-ceph-osd_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:45.791 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:45.837 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-google-auth. 2026-03-21T06:56:45.843 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../57-python3-google-auth_1.5.1-3_all.deb ... 2026-03-21T06:56:45.845 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-21T06:56:45.892 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-21T06:56:45.896 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../58-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-21T06:56:45.900 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-21T06:56:45.905 INFO:teuthology.orchestra.run.vm04.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 multipath-tools amd64 0.8.8-1ubuntu1.22.04.4 [331 kB] 2026-03-21T06:56:45.939 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-websocket. 2026-03-21T06:56:45.945 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../59-python3-websocket_1.2.3-1_all.deb ... 2026-03-21T06:56:45.950 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-21T06:56:45.995 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-21T06:56:46.001 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../60-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-21T06:56:46.007 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-21T06:56:46.263 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-21T06:56:46.270 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../61-ceph-mgr-k8sevents_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:46.280 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:46.293 INFO:teuthology.orchestra.run.vm04.stdout:Fetched 486 kB in 1s (593 kB/s) 2026-03-21T06:56:46.303 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph. 2026-03-21T06:56:46.309 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../41-ceph_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:46.391 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:46.397 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-jmespath. 2026-03-21T06:56:46.399 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-21T06:56:46.404 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../62-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-21T06:56:46.419 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-21T06:56:46.427 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126082 files and directories currently installed.) 2026-03-21T06:56:46.429 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-21T06:56:46.433 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-21T06:56:46.433 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-fuse. 2026-03-21T06:56:46.439 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../42-ceph-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:46.445 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:46.456 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-21T06:56:46.461 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../63-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-21T06:56:46.472 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:46.475 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-21T06:56:46.482 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-21T06:56:46.490 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-21T06:56:46.512 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package jq. 2026-03-21T06:56:46.517 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../64-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-21T06:56:46.519 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-mds. 2026-03-21T06:56:46.520 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:46.525 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../43-ceph-mds_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:46.531 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package s3cmd. 2026-03-21T06:56:46.531 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:46.538 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../archives/s3cmd_2.2.0-1_all.deb ... 2026-03-21T06:56:46.543 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking s3cmd (2.2.0-1) ... 2026-03-21T06:56:46.576 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package socat. 2026-03-21T06:56:46.580 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../65-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-21T06:56:46.597 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-21T06:56:46.632 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package cephadm. 2026-03-21T06:56:46.634 INFO:teuthology.orchestra.run.vm04.stdout:Selecting previously unselected package multipath-tools. 2026-03-21T06:56:46.638 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../44-cephadm_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:46.639 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:46.641 INFO:teuthology.orchestra.run.vm04.stdout:Preparing to unpack .../multipath-tools_0.8.8-1ubuntu1.22.04.4_amd64.deb ... 2026-03-21T06:56:46.649 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package xmlstarlet. 2026-03-21T06:56:46.652 INFO:teuthology.orchestra.run.vm04.stdout:Unpacking multipath-tools (0.8.8-1ubuntu1.22.04.4) ... 2026-03-21T06:56:46.655 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../66-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-21T06:56:46.660 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-21T06:56:46.698 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-asyncssh. 2026-03-21T06:56:46.704 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../45-python3-asyncssh_2.5.0-1ubuntu0.1_all.deb ... 2026-03-21T06:56:46.710 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:46.723 INFO:teuthology.orchestra.run.vm04.stdout:Setting up s3cmd (2.2.0-1) ... 2026-03-21T06:56:46.742 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-test. 2026-03-21T06:56:46.748 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../67-ceph-test_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:46.749 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:46.756 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-mgr-cephadm. 2026-03-21T06:56:46.762 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../46-ceph-mgr-cephadm_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:46.778 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:46.823 INFO:teuthology.orchestra.run.vm04.stdout:Setting up multipath-tools (0.8.8-1ubuntu1.22.04.4) ... 2026-03-21T06:56:46.852 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-repoze.lru. 2026-03-21T06:56:46.859 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../47-python3-repoze.lru_0.7-2_all.deb ... 2026-03-21T06:56:46.862 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-repoze.lru (0.7-2) ... 2026-03-21T06:56:46.911 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-routes. 2026-03-21T06:56:46.917 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../48-python3-routes_2.5.1-1ubuntu1_all.deb ... 2026-03-21T06:56:46.935 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-routes (2.5.1-1ubuntu1) ... 2026-03-21T06:56:46.972 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:46 vm04 bash[20194]: cluster 2026-03-21T06:56:45.612938+0000 mgr.x (mgr.14152) 390 : cluster [DBG] pgmap v317: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:46.972 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:46 vm04 bash[20194]: cluster 2026-03-21T06:56:45.612938+0000 mgr.x (mgr.14152) 390 : cluster [DBG] pgmap v317: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:46.995 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-mgr-dashboard. 2026-03-21T06:56:47.001 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../49-ceph-mgr-dashboard_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:47.006 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:47.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:46 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:56:47.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:46 vm02 bash[17657]: cluster 2026-03-21T06:56:45.612938+0000 mgr.x (mgr.14152) 390 : cluster [DBG] pgmap v317: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:47.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:46 vm02 bash[17657]: cluster 2026-03-21T06:56:45.612938+0000 mgr.x (mgr.14152) 390 : cluster [DBG] pgmap v317: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:47.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:46 vm07 bash[19945]: cluster 2026-03-21T06:56:45.612938+0000 mgr.x (mgr.14152) 390 : cluster [DBG] pgmap v317: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:47.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:46 vm07 bash[19945]: cluster 2026-03-21T06:56:45.612938+0000 mgr.x (mgr.14152) 390 : cluster [DBG] pgmap v317: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:47.223 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:46 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:47.223 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:46 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:47.223 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:46 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:47.223 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 06:56:47 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:47.223 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:46 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:47.333 INFO:teuthology.orchestra.run.vm04.stdout:Could not execute systemctl: at /usr/bin/deb-systemd-invoke line 142. 2026-03-21T06:56:47.342 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-21T06:56:47.450 INFO:teuthology.orchestra.run.vm04.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-21T06:56:47.508 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:47 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:47.508 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 06:56:47 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:47.509 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 06:56:47 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:47.555 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T06:56:47.687 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T06:56:47.901 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:47 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:56:48.364 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-sklearn-lib:amd64. 2026-03-21T06:56:48.371 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../50-python3-sklearn-lib_0.23.2-5ubuntu6_amd64.deb ... 2026-03-21T06:56:48.377 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-21T06:56:48.513 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-joblib. 2026-03-21T06:56:48.518 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:48.518 INFO:teuthology.orchestra.run.vm04.stdout:Running kernel seems to be up-to-date. 2026-03-21T06:56:48.518 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:48.518 INFO:teuthology.orchestra.run.vm04.stdout:Services to be restarted: 2026-03-21T06:56:48.520 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../51-python3-joblib_0.17.0-4ubuntu1_all.deb ... 2026-03-21T06:56:48.521 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-joblib (0.17.0-4ubuntu1) ... 2026-03-21T06:56:48.522 INFO:teuthology.orchestra.run.vm04.stdout: systemctl restart apache-htcacheclean.service 2026-03-21T06:56:48.529 INFO:teuthology.orchestra.run.vm04.stdout: systemctl restart rsyslog.service 2026-03-21T06:56:48.533 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:48.533 INFO:teuthology.orchestra.run.vm04.stdout:Service restarts being deferred: 2026-03-21T06:56:48.534 INFO:teuthology.orchestra.run.vm04.stdout: systemctl restart networkd-dispatcher.service 2026-03-21T06:56:48.534 INFO:teuthology.orchestra.run.vm04.stdout: systemctl restart unattended-upgrades.service 2026-03-21T06:56:48.534 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:48.534 INFO:teuthology.orchestra.run.vm04.stdout:No containers need to be restarted. 2026-03-21T06:56:48.534 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:48.534 INFO:teuthology.orchestra.run.vm04.stdout:No user sessions are running outdated binaries. 2026-03-21T06:56:48.534 INFO:teuthology.orchestra.run.vm04.stdout: 2026-03-21T06:56:48.534 INFO:teuthology.orchestra.run.vm04.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-21T06:56:48.587 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-threadpoolctl. 2026-03-21T06:56:48.594 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../52-python3-threadpoolctl_3.1.0-1_all.deb ... 2026-03-21T06:56:48.604 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-threadpoolctl (3.1.0-1) ... 2026-03-21T06:56:48.652 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-sklearn. 2026-03-21T06:56:48.659 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../53-python3-sklearn_0.23.2-5ubuntu6_all.deb ... 2026-03-21T06:56:48.663 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-21T06:56:48.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:48 vm04 bash[20194]: audit 2026-03-21T06:56:46.663459+0000 mgr.x (mgr.14152) 391 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:48.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:48 vm04 bash[20194]: audit 2026-03-21T06:56:46.663459+0000 mgr.x (mgr.14152) 391 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:48.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:48 vm02 bash[17657]: audit 2026-03-21T06:56:46.663459+0000 mgr.x (mgr.14152) 391 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:48.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:48 vm02 bash[17657]: audit 2026-03-21T06:56:46.663459+0000 mgr.x (mgr.14152) 391 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:48 vm07 bash[19945]: audit 2026-03-21T06:56:46.663459+0000 mgr.x (mgr.14152) 391 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:48.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:48 vm07 bash[19945]: audit 2026-03-21T06:56:46.663459+0000 mgr.x (mgr.14152) 391 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:48.965 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-mgr-diskprediction-local. 2026-03-21T06:56:48.972 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../54-ceph-mgr-diskprediction-local_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:48.973 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:48.986 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package ceph-volume. 2026-03-21T06:56:48.992 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../68-ceph-volume_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:48.993 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:49.023 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libcephfs-daemon. 2026-03-21T06:56:49.030 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../69-libcephfs-daemon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:49.031 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:49.050 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libcephfs-proxy2. 2026-03-21T06:56:49.057 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../70-libcephfs-proxy2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:49.058 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:49.076 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-21T06:56:49.081 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../71-libcephfs-dev_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:49.082 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:49.102 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package nvme-cli. 2026-03-21T06:56:49.109 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../72-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-21T06:56:49.111 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-21T06:56:49.149 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-21T06:56:49.156 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../73-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-21T06:56:49.156 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:49.224 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-21T06:56:49.230 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../74-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-21T06:56:49.232 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-21T06:56:49.248 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pluggy. 2026-03-21T06:56:49.254 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../75-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-21T06:56:49.255 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-21T06:56:49.272 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-psutil. 2026-03-21T06:56:49.279 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../76-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-21T06:56:49.280 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-21T06:56:49.379 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-py. 2026-03-21T06:56:49.385 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../77-python3-py_1.10.0-1_all.deb ... 2026-03-21T06:56:49.385 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-cachetools. 2026-03-21T06:56:49.386 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-21T06:56:49.391 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../55-python3-cachetools_5.0.0-1_all.deb ... 2026-03-21T06:56:49.392 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-cachetools (5.0.0-1) ... 2026-03-21T06:56:49.406 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-rsa. 2026-03-21T06:56:49.406 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T06:56:49.408 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pygments. 2026-03-21T06:56:49.411 DEBUG:teuthology.parallel:result is None 2026-03-21T06:56:49.411 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../56-python3-rsa_4.8-1_all.deb ... 2026-03-21T06:56:49.412 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-rsa (4.8-1) ... 2026-03-21T06:56:49.414 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../78-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-21T06:56:49.415 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-21T06:56:49.431 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-google-auth. 2026-03-21T06:56:49.436 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../57-python3-google-auth_1.5.1-3_all.deb ... 2026-03-21T06:56:49.437 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-google-auth (1.5.1-3) ... 2026-03-21T06:56:49.455 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-requests-oauthlib. 2026-03-21T06:56:49.460 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../58-python3-requests-oauthlib_1.3.0+ds-0.1_all.deb ... 2026-03-21T06:56:49.462 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-21T06:56:49.477 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-websocket. 2026-03-21T06:56:49.482 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../59-python3-websocket_1.2.3-1_all.deb ... 2026-03-21T06:56:49.482 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-websocket (1.2.3-1) ... 2026-03-21T06:56:49.485 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-toml. 2026-03-21T06:56:49.493 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../79-python3-toml_0.10.2-1_all.deb ... 2026-03-21T06:56:49.494 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-21T06:56:49.501 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-kubernetes. 2026-03-21T06:56:49.505 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../60-python3-kubernetes_12.0.1-1ubuntu1_all.deb ... 2026-03-21T06:56:49.506 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-21T06:56:49.511 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-pytest. 2026-03-21T06:56:49.518 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../80-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-21T06:56:49.518 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-21T06:56:49.568 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-simplejson. 2026-03-21T06:56:49.573 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../81-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-21T06:56:49.573 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-21T06:56:49.596 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-webob. 2026-03-21T06:56:49.601 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../82-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-21T06:56:49.601 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-21T06:56:49.624 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-21T06:56:49.628 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../83-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-21T06:56:49.629 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-21T06:56:49.716 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-mgr-k8sevents. 2026-03-21T06:56:49.724 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../61-ceph-mgr-k8sevents_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:49.725 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:49.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:49 vm04 bash[20194]: audit 2026-03-21T06:56:47.498063+0000 mgr.x (mgr.14152) 392 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:49.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:49 vm04 bash[20194]: audit 2026-03-21T06:56:47.498063+0000 mgr.x (mgr.14152) 392 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:49.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:49 vm04 bash[20194]: cluster 2026-03-21T06:56:47.613367+0000 mgr.x (mgr.14152) 393 : cluster [DBG] pgmap v318: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:49.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:49 vm04 bash[20194]: cluster 2026-03-21T06:56:47.613367+0000 mgr.x (mgr.14152) 393 : cluster [DBG] pgmap v318: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:49.760 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libonig5:amd64. 2026-03-21T06:56:49.766 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../62-libonig5_6.9.7.1-2build1_amd64.deb ... 2026-03-21T06:56:49.767 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-21T06:56:49.784 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package radosgw. 2026-03-21T06:56:49.785 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libjq1:amd64. 2026-03-21T06:56:49.790 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../84-radosgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:49.790 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../63-libjq1_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-21T06:56:49.791 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:49.791 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:49.807 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package jq. 2026-03-21T06:56:49.812 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../64-jq_1.6-2.1ubuntu3.1_amd64.deb ... 2026-03-21T06:56:49.813 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking jq (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:49.827 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package socat. 2026-03-21T06:56:49.833 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../65-socat_1.7.4.1-3ubuntu4_amd64.deb ... 2026-03-21T06:56:49.834 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking socat (1.7.4.1-3ubuntu4) ... 2026-03-21T06:56:49.857 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package xmlstarlet. 2026-03-21T06:56:49.863 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../66-xmlstarlet_1.6.1-2.1_amd64.deb ... 2026-03-21T06:56:49.864 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking xmlstarlet (1.6.1-2.1) ... 2026-03-21T06:56:49.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:49 vm02 bash[17657]: audit 2026-03-21T06:56:47.498063+0000 mgr.x (mgr.14152) 392 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:49.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:49 vm02 bash[17657]: audit 2026-03-21T06:56:47.498063+0000 mgr.x (mgr.14152) 392 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:49.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:49 vm02 bash[17657]: cluster 2026-03-21T06:56:47.613367+0000 mgr.x (mgr.14152) 393 : cluster [DBG] pgmap v318: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:49.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:49 vm02 bash[17657]: cluster 2026-03-21T06:56:47.613367+0000 mgr.x (mgr.14152) 393 : cluster [DBG] pgmap v318: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:49 vm07 bash[19945]: audit 2026-03-21T06:56:47.498063+0000 mgr.x (mgr.14152) 392 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:49 vm07 bash[19945]: audit 2026-03-21T06:56:47.498063+0000 mgr.x (mgr.14152) 392 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:49 vm07 bash[19945]: cluster 2026-03-21T06:56:47.613367+0000 mgr.x (mgr.14152) 393 : cluster [DBG] pgmap v318: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:49.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:49 vm07 bash[19945]: cluster 2026-03-21T06:56:47.613367+0000 mgr.x (mgr.14152) 393 : cluster [DBG] pgmap v318: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:49.909 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-test. 2026-03-21T06:56:49.914 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../67-ceph-test_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:49.915 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:50.340 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package rbd-fuse. 2026-03-21T06:56:50.345 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../85-rbd-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:50.346 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:50.365 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package smartmontools. 2026-03-21T06:56:50.371 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../86-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-21T06:56:50.380 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-21T06:56:50.425 INFO:teuthology.orchestra.run.vm02.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-21T06:56:50.695 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-21T06:56:50.695 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-21T06:56:50.695 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:50 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:50.695 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:50 vm02 bash[17657]: cluster 2026-03-21T06:56:49.613816+0000 mgr.x (mgr.14152) 394 : cluster [DBG] pgmap v319: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:50.695 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:50 vm02 bash[17657]: cluster 2026-03-21T06:56:49.613816+0000 mgr.x (mgr.14152) 394 : cluster [DBG] pgmap v319: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:50.695 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:50 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:50.695 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:50 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:50.695 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:50 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:50.695 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:50 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:50.998 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:50 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:50.998 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:50 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:50.998 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:50 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:50.998 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:50 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:50.998 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:50 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:50 vm04 bash[20194]: cluster 2026-03-21T06:56:49.613816+0000 mgr.x (mgr.14152) 394 : cluster [DBG] pgmap v319: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:50 vm04 bash[20194]: cluster 2026-03-21T06:56:49.613816+0000 mgr.x (mgr.14152) 394 : cluster [DBG] pgmap v319: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:51.098 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-21T06:56:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:50 vm07 bash[19945]: cluster 2026-03-21T06:56:49.613816+0000 mgr.x (mgr.14152) 394 : cluster [DBG] pgmap v319: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:50 vm07 bash[19945]: cluster 2026-03-21T06:56:49.613816+0000 mgr.x (mgr.14152) 394 : cluster [DBG] pgmap v319: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:51.229 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-21T06:56:51.233 INFO:teuthology.orchestra.run.vm02.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-21T06:56:51.298 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-21T06:56:51.298 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.298 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.298 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.298 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.299 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.551 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-21T06:56:51.551 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.551 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.551 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.551 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.551 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.798 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package ceph-volume. 2026-03-21T06:56:51.804 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../68-ceph-volume_20.2.0-712-g70f8415b-1jammy_all.deb ... 2026-03-21T06:56:51.805 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:51.832 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libcephfs-daemon. 2026-03-21T06:56:51.838 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../69-libcephfs-daemon_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:51.839 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:51.854 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libcephfs-proxy2. 2026-03-21T06:56:51.860 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../70-libcephfs-proxy2_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:51.861 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:51.875 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package libcephfs-dev. 2026-03-21T06:56:51.881 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../71-libcephfs-dev_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:51.881 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:51.887 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.887 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.888 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.888 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:51.899 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package nvme-cli. 2026-03-21T06:56:51.905 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../72-nvme-cli_1.16-3ubuntu0.3_amd64.deb ... 2026-03-21T06:56:51.906 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking nvme-cli (1.16-3ubuntu0.3) ... 2026-03-21T06:56:51.944 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python-asyncssh-doc. 2026-03-21T06:56:51.950 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../73-python-asyncssh-doc_2.5.0-1ubuntu0.1_all.deb ... 2026-03-21T06:56:51.951 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:51.999 INFO:teuthology.orchestra.run.vm02.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-21T06:56:52.001 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-iniconfig. 2026-03-21T06:56:52.008 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../74-python3-iniconfig_1.1.1-2_all.deb ... 2026-03-21T06:56:52.009 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-iniconfig (1.1.1-2) ... 2026-03-21T06:56:52.016 INFO:teuthology.orchestra.run.vm02.stdout:Setting up cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:52.027 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-pluggy. 2026-03-21T06:56:52.034 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../75-python3-pluggy_0.13.0-7.1_all.deb ... 2026-03-21T06:56:52.035 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-pluggy (0.13.0-7.1) ... 2026-03-21T06:56:52.057 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-psutil. 2026-03-21T06:56:52.061 INFO:teuthology.orchestra.run.vm02.stdout:Adding system user cephadm....done 2026-03-21T06:56:52.064 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../76-python3-psutil_5.9.0-1build1_amd64.deb ... 2026-03-21T06:56:52.065 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-psutil (5.9.0-1build1) ... 2026-03-21T06:56:52.071 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-21T06:56:52.093 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-py. 2026-03-21T06:56:52.099 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../77-python3-py_1.10.0-1_all.deb ... 2026-03-21T06:56:52.100 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-py (1.10.0-1) ... 2026-03-21T06:56:52.126 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-pygments. 2026-03-21T06:56:52.132 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../78-python3-pygments_2.11.2+dfsg-2ubuntu0.1_all.deb ... 2026-03-21T06:56:52.133 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-21T06:56:52.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:52.138 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:52.138 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:52.138 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:52.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:51 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:52.139 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:52.141 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-21T06:56:52.197 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-toml. 2026-03-21T06:56:52.203 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../79-python3-toml_0.10.2-1_all.deb ... 2026-03-21T06:56:52.204 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-toml (0.10.2-1) ... 2026-03-21T06:56:52.208 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-21T06:56:52.221 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-pytest. 2026-03-21T06:56:52.228 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../80-python3-pytest_6.2.5-1ubuntu2_all.deb ... 2026-03-21T06:56:52.229 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-pytest (6.2.5-1ubuntu2) ... 2026-03-21T06:56:52.271 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-simplejson. 2026-03-21T06:56:52.275 INFO:teuthology.orchestra.run.vm02.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-21T06:56:52.277 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-21T06:56:52.277 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../81-python3-simplejson_3.17.6-1build1_amd64.deb ... 2026-03-21T06:56:52.278 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-simplejson (3.17.6-1build1) ... 2026-03-21T06:56:52.300 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-webob. 2026-03-21T06:56:52.305 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../82-python3-webob_1%3a1.8.6-1.1ubuntu0.1_all.deb ... 2026-03-21T06:56:52.306 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-21T06:56:52.326 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package qttranslations5-l10n. 2026-03-21T06:56:52.331 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../83-qttranslations5-l10n_5.15.3-1_all.deb ... 2026-03-21T06:56:52.332 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking qttranslations5-l10n (5.15.3-1) ... 2026-03-21T06:56:52.365 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-21T06:56:52.462 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package radosgw. 2026-03-21T06:56:52.468 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../84-radosgw_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:52.469 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:52.483 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-21T06:56:52.549 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-21T06:56:52.616 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:52.686 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-21T06:56:52.689 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-21T06:56:52.691 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-21T06:56:52.693 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-21T06:56:52.695 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-21T06:56:52.813 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-21T06:56:52.882 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:52.884 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-21T06:56:52.956 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-21T06:56:53.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:52 vm04 bash[20194]: cluster 2026-03-21T06:56:51.614117+0000 mgr.x (mgr.14152) 395 : cluster [DBG] pgmap v320: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:53.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:52 vm04 bash[20194]: cluster 2026-03-21T06:56:51.614117+0000 mgr.x (mgr.14152) 395 : cluster [DBG] pgmap v320: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:53.069 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-21T06:56:53.084 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package rbd-fuse. 2026-03-21T06:56:53.090 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../85-rbd-fuse_20.2.0-712-g70f8415b-1jammy_amd64.deb ... 2026-03-21T06:56:53.091 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:53.114 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package smartmontools. 2026-03-21T06:56:53.121 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../86-smartmontools_7.2-1ubuntu0.1_amd64.deb ... 2026-03-21T06:56:53.131 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking smartmontools (7.2-1ubuntu0.1) ... 2026-03-21T06:56:53.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:52 vm02 bash[17657]: cluster 2026-03-21T06:56:51.614117+0000 mgr.x (mgr.14152) 395 : cluster [DBG] pgmap v320: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:53.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:52 vm02 bash[17657]: cluster 2026-03-21T06:56:51.614117+0000 mgr.x (mgr.14152) 395 : cluster [DBG] pgmap v320: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:52 vm07 bash[19945]: cluster 2026-03-21T06:56:51.614117+0000 mgr.x (mgr.14152) 395 : cluster [DBG] pgmap v320: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:52 vm07 bash[19945]: cluster 2026-03-21T06:56:51.614117+0000 mgr.x (mgr.14152) 395 : cluster [DBG] pgmap v320: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:53.175 INFO:teuthology.orchestra.run.vm07.stdout:Setting up smartmontools (7.2-1ubuntu0.1) ... 2026-03-21T06:56:53.339 INFO:teuthology.orchestra.run.vm02.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-21T06:56:53.341 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-21T06:56:53.439 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/smartd.service → /lib/systemd/system/smartmontools.service. 2026-03-21T06:56:53.439 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/smartmontools.service → /lib/systemd/system/smartmontools.service. 2026-03-21T06:56:53.445 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:53.551 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.551 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.551 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.551 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.551 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.584 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-21T06:56:53.694 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-21T06:56:53.799 INFO:teuthology.orchestra.run.vm02.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-21T06:56:53.801 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:53.844 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.845 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.845 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.845 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.845 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.845 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.845 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.845 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:53 vm07 bash[19945]: cluster 2026-03-21T06:56:53.614435+0000 mgr.x (mgr.14152) 396 : cluster [DBG] pgmap v321: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:53.845 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:53 vm07 bash[19945]: cluster 2026-03-21T06:56:53.614435+0000 mgr.x (mgr.14152) 396 : cluster [DBG] pgmap v321: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:53.845 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.845 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.845 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:53 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:53.847 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-iniconfig (1.1.1-2) ... 2026-03-21T06:56:53.893 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-21T06:56:53.911 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-21T06:56:53.914 INFO:teuthology.orchestra.run.vm07.stdout:Setting up nvme-cli (1.16-3ubuntu0.3) ... 2026-03-21T06:56:53.977 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmefc-boot-connections.service → /lib/systemd/system/nvmefc-boot-connections.service. 2026-03-21T06:56:54.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:53 vm04 bash[20194]: cluster 2026-03-21T06:56:53.614435+0000 mgr.x (mgr.14152) 396 : cluster [DBG] pgmap v321: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:54.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:53 vm04 bash[20194]: cluster 2026-03-21T06:56:53.614435+0000 mgr.x (mgr.14152) 396 : cluster [DBG] pgmap v321: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:54.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:53 vm02 bash[17657]: cluster 2026-03-21T06:56:53.614435+0000 mgr.x (mgr.14152) 396 : cluster [DBG] pgmap v321: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:54.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:53 vm02 bash[17657]: cluster 2026-03-21T06:56:53.614435+0000 mgr.x (mgr.14152) 396 : cluster [DBG] pgmap v321: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.151 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.151 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.151 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.151 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.221 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/default.target.wants/nvmf-autoconnect.service → /lib/systemd/system/nvmf-autoconnect.service. 2026-03-21T06:56:54.442 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:54.447 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-21T06:56:54.519 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-21T06:56:54.521 INFO:teuthology.orchestra.run.vm02.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-21T06:56:54.523 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-21T06:56:54.564 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.564 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.564 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.564 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.564 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.600 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-21T06:56:54.667 INFO:teuthology.orchestra.run.vm07.stdout:nvmf-connect.target is a disabled or a static unit, not starting it. 2026-03-21T06:56:54.669 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:54.671 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-21T06:56:54.683 INFO:teuthology.orchestra.run.vm07.stdout:Setting up cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:54.729 INFO:teuthology.orchestra.run.vm07.stdout:Adding system user cephadm....done 2026-03-21T06:56:54.738 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-jaraco.classes (3.2.1-3) ... 2026-03-21T06:56:54.744 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-21T06:56:54.804 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:54.806 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-jaraco.functools (3.4.0-2) ... 2026-03-21T06:56:54.813 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-21T06:56:54.871 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-repoze.lru (0.7-2) ... 2026-03-21T06:56:54.883 INFO:teuthology.orchestra.run.vm02.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-21T06:56:54.884 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-21T06:56:54.901 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.901 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.901 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.901 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:54 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:54.943 INFO:teuthology.orchestra.run.vm07.stdout:Setting up liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-21T06:56:54.946 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-py (1.10.0-1) ... 2026-03-21T06:56:54.959 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-21T06:56:54.961 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-21T06:56:55.028 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-21T06:56:55.038 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-joblib (0.17.0-4ubuntu1) ... 2026-03-21T06:56:55.115 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-21T06:56:55.182 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:55.184 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-21T06:56:55.220 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-cachetools (5.0.0-1) ... 2026-03-21T06:56:55.290 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-threadpoolctl (3.1.0-1) ... 2026-03-21T06:56:55.315 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-21T06:56:55.358 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:55.383 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:55.385 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-21T06:56:55.428 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-21T06:56:55.431 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libnbd0 (1.10.5-1) ... 2026-03-21T06:56:55.433 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-21T06:56:55.435 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-21T06:56:55.437 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-psutil (5.9.0-1build1) ... 2026-03-21T06:56:55.463 INFO:teuthology.orchestra.run.vm02.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:55.465 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-21T06:56:55.559 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-natsort (8.0.2-1) ... 2026-03-21T06:56:55.593 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-21T06:56:55.595 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librados2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:55.597 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:55.599 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:55.601 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-21T06:56:55.630 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:55.632 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-routes (2.5.1-1ubuntu1) ... 2026-03-21T06:56:55.707 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-simplejson (3.17.6-1build1) ... 2026-03-21T06:56:55.790 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-pygments (2.11.2+dfsg-2ubuntu0.1) ... 2026-03-21T06:56:56.078 INFO:teuthology.orchestra.run.vm07.stdout:Setting up qttranslations5-l10n (5.15.3-1) ... 2026-03-21T06:56:56.081 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-21T06:56:56.199 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.201 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.203 INFO:teuthology.orchestra.run.vm02.stdout:Setting up librbd1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.205 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.207 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.257 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-21T06:56:56.270 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-21T06:56:56.270 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-21T06:56:56.399 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-21T06:56:56.493 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-jaraco.text (3.6.0-2) ... 2026-03-21T06:56:56.564 INFO:teuthology.orchestra.run.vm07.stdout:Setting up socat (1.7.4.1-3ubuntu4) ... 2026-03-21T06:56:56.567 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.591 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:56.591 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:56.591 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:56.592 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:56.592 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:56.668 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-21T06:56:56.684 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.686 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.688 INFO:teuthology.orchestra.run.vm02.stdout:Setting up libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.690 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.692 INFO:teuthology.orchestra.run.vm02.stdout:Setting up rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.694 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.696 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.698 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:56.729 INFO:teuthology.orchestra.run.vm02.stdout:Adding group ceph....done 2026-03-21T06:56:56.765 INFO:teuthology.orchestra.run.vm02.stdout:Adding system user ceph....done 2026-03-21T06:56:56.774 INFO:teuthology.orchestra.run.vm02.stdout:Setting system user ceph properties....done 2026-03-21T06:56:56.779 INFO:teuthology.orchestra.run.vm02.stdout:Fixing /var/run/ceph ownership....done 2026-03-21T06:56:56.783 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:56.783 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:56.783 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:56.783 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:56 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:56:56.783 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:56.784 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:56.784 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:56 vm02 bash[17657]: cluster 2026-03-21T06:56:55.614768+0000 mgr.x (mgr.14152) 397 : cluster [DBG] pgmap v322: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:56.784 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:56 vm02 bash[17657]: cluster 2026-03-21T06:56:55.614768+0000 mgr.x (mgr.14152) 397 : cluster [DBG] pgmap v322: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:57.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:56 vm04 bash[20194]: cluster 2026-03-21T06:56:55.614768+0000 mgr.x (mgr.14152) 397 : cluster [DBG] pgmap v322: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:57.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:56 vm04 bash[20194]: cluster 2026-03-21T06:56:55.614768+0000 mgr.x (mgr.14152) 397 : cluster [DBG] pgmap v322: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:57.101 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-21T06:56:57.101 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.101 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.101 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.101 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.101 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:56 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:56 vm07 bash[19945]: cluster 2026-03-21T06:56:55.614768+0000 mgr.x (mgr.14152) 397 : cluster [DBG] pgmap v322: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:57.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:56 vm07 bash[19945]: cluster 2026-03-21T06:56:55.614768+0000 mgr.x (mgr.14152) 397 : cluster [DBG] pgmap v322: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:56:57.235 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:57.241 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-toml (0.10.2-1) ... 2026-03-21T06:56:57.312 INFO:teuthology.orchestra.run.vm07.stdout:Setting up librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-21T06:56:57.314 INFO:teuthology.orchestra.run.vm07.stdout:Setting up xmlstarlet (1.6.1-2.1) ... 2026-03-21T06:56:57.317 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-pluggy (0.13.0-7.1) ... 2026-03-21T06:56:57.388 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.388 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.388 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.388 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.388 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.390 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-zc.lockfile (2.0-1) ... 2026-03-21T06:56:57.464 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:57.466 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-rsa (4.8-1) ... 2026-03-21T06:56:57.542 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-tempora (4.1.2-1) ... 2026-03-21T06:56:57.557 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:57.559 INFO:teuthology.orchestra.run.vm02.stdout:Setting up radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:57.612 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-prettytable (2.5.0-2) ... 2026-03-21T06:56:57.686 INFO:teuthology.orchestra.run.vm07.stdout:Setting up liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-21T06:56:57.688 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-websocket (1.2.3-1) ... 2026-03-21T06:56:57.717 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.718 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.718 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.718 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:57 vm02 bash[17657]: audit 2026-03-21T06:56:56.667387+0000 mgr.x (mgr.14152) 398 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:57.718 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:57 vm02 bash[17657]: audit 2026-03-21T06:56:56.667387+0000 mgr.x (mgr.14152) 398 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:57.718 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.718 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.770 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-21T06:56:57.772 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-21T06:56:57.845 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-21T06:56:57.856 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-21T06:56:57.856 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-21T06:56:57.901 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:57 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:56:57.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:57 vm07 bash[19945]: audit 2026-03-21T06:56:56.667387+0000 mgr.x (mgr.14152) 398 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:57.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:57 vm07 bash[19945]: audit 2026-03-21T06:56:56.667387+0000 mgr.x (mgr.14152) 398 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:57.936 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-jaraco.collections (3.4.0-2) ... 2026-03-21T06:56:57.968 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.968 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.968 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.969 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:57.969 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:57 vm04 bash[20194]: audit 2026-03-21T06:56:56.667387+0000 mgr.x (mgr.14152) 398 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:58.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:57 vm04 bash[20194]: audit 2026-03-21T06:56:56.667387+0000 mgr.x (mgr.14152) 398 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:58.009 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:58.011 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-pytest (6.2.5-1ubuntu2) ... 2026-03-21T06:56:58.148 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-portend (3.0.0-1) ... 2026-03-21T06:56:58.230 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T06:56:58.232 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-google-auth (1.5.1-3) ... 2026-03-21T06:56:58.237 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.237 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.237 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.237 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.237 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.237 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.237 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.237 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.237 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:57 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.238 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.239 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:58.308 INFO:teuthology.orchestra.run.vm07.stdout:Setting up jq (1.6-2.1ubuntu3.1) ... 2026-03-21T06:56:58.311 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-cherrypy3 (18.6.1-4) ... 2026-03-21T06:56:58.332 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-21T06:56:58.459 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-21T06:56:58.462 INFO:teuthology.orchestra.run.vm07.stdout:Setting up librados2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:58.464 INFO:teuthology.orchestra.run.vm07.stdout:Setting up librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:58.466 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:58.468 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-21T06:56:58.539 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.539 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.540 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.540 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.540 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:58 vm02 bash[17657]: audit 2026-03-21T06:56:57.505804+0000 mgr.x (mgr.14152) 399 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:58.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:58 vm02 bash[17657]: audit 2026-03-21T06:56:57.505804+0000 mgr.x (mgr.14152) 399 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:58.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:58 vm02 bash[17657]: cluster 2026-03-21T06:56:57.615057+0000 mgr.x (mgr.14152) 400 : cluster [DBG] pgmap v323: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:58.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:58 vm02 bash[17657]: cluster 2026-03-21T06:56:57.615057+0000 mgr.x (mgr.14152) 400 : cluster [DBG] pgmap v323: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:58.793 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.793 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.793 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.793 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.793 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:58.801 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:58.863 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-21T06:56:58.863 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-21T06:56:59.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:58 vm04 bash[20194]: audit 2026-03-21T06:56:57.505804+0000 mgr.x (mgr.14152) 399 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:59.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:58 vm04 bash[20194]: audit 2026-03-21T06:56:57.505804+0000 mgr.x (mgr.14152) 399 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:59.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:58 vm04 bash[20194]: cluster 2026-03-21T06:56:57.615057+0000 mgr.x (mgr.14152) 400 : cluster [DBG] pgmap v323: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:59.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:58 vm04 bash[20194]: cluster 2026-03-21T06:56:57.615057+0000 mgr.x (mgr.14152) 400 : cluster [DBG] pgmap v323: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:59.053 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.053 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.053 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.053 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.053 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:58 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.090 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.092 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.094 INFO:teuthology.orchestra.run.vm07.stdout:Setting up librbd1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.096 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.098 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:58 vm07 bash[19945]: audit 2026-03-21T06:56:57.505804+0000 mgr.x (mgr.14152) 399 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:58 vm07 bash[19945]: audit 2026-03-21T06:56:57.505804+0000 mgr.x (mgr.14152) 399 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:56:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:58 vm07 bash[19945]: cluster 2026-03-21T06:56:57.615057+0000 mgr.x (mgr.14152) 400 : cluster [DBG] pgmap v323: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:58 vm07 bash[19945]: cluster 2026-03-21T06:56:57.615057+0000 mgr.x (mgr.14152) 400 : cluster [DBG] pgmap v323: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:56:59.156 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/remote-fs.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-21T06:56:59.156 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-fuse.target → /lib/systemd/system/ceph-fuse.target. 2026-03-21T06:56:59.248 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.319 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-21T06:56:59.319 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-21T06:56:59.319 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.319 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.319 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.320 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.320 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.457 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.457 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.457 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.457 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.457 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.553 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.555 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.557 INFO:teuthology.orchestra.run.vm07.stdout:Setting up libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.560 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.563 INFO:teuthology.orchestra.run.vm07.stdout:Setting up rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.565 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.567 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.569 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.601 INFO:teuthology.orchestra.run.vm07.stdout:Adding group ceph....done 2026-03-21T06:56:59.627 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.627 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.627 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.628 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.628 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.636 INFO:teuthology.orchestra.run.vm07.stdout:Adding system user ceph....done 2026-03-21T06:56:59.644 INFO:teuthology.orchestra.run.vm07.stdout:Setting system user ceph properties....done 2026-03-21T06:56:59.649 INFO:teuthology.orchestra.run.vm07.stdout:Fixing /var/run/ceph ownership....done 2026-03-21T06:56:59.652 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.652 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.652 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.653 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.653 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.717 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:59 vm07 bash[19945]: audit 2026-03-21T06:56:59.054233+0000 mon.a (mon.0) 853 : audit [DBG] from='client.? 192.168.123.102:0/811624370' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:56:59.717 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:59 vm07 bash[19945]: audit 2026-03-21T06:56:59.054233+0000 mon.a (mon.0) 853 : audit [DBG] from='client.? 192.168.123.102:0/811624370' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:56:59.721 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:56:59.803 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-21T06:56:59.803 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-21T06:56:59.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:59 vm02 bash[17657]: audit 2026-03-21T06:56:59.054233+0000 mon.a (mon.0) 853 : audit [DBG] from='client.? 192.168.123.102:0/811624370' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:56:59.888 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:59 vm02 bash[17657]: audit 2026-03-21T06:56:59.054233+0000 mon.a (mon.0) 853 : audit [DBG] from='client.? 192.168.123.102:0/811624370' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:56:59.888 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.888 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.889 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.968 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/rbdmap.service → /lib/systemd/system/rbdmap.service. 2026-03-21T06:56:59.968 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.968 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.968 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.968 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:56:59.968 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:56:59 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:59 vm04 bash[20194]: audit 2026-03-21T06:56:59.054233+0000 mon.a (mon.0) 853 : audit [DBG] from='client.? 192.168.123.102:0/811624370' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:57:00.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:56:59 vm04 bash[20194]: audit 2026-03-21T06:56:59.054233+0000 mon.a (mon.0) 853 : audit [DBG] from='client.? 192.168.123.102:0/811624370' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:57:00.200 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.201 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.201 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.201 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.201 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.201 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.201 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.201 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.201 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:56:59 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.201 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.203 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:00.205 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:00.217 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:00.263 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.263 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.263 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.264 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.264 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.281 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-21T06:57:00.281 INFO:teuthology.orchestra.run.vm02.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-21T06:57:00.374 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:00.376 INFO:teuthology.orchestra.run.vm07.stdout:Setting up radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:00.467 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.467 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.467 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.467 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.468 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.516 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.516 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.516 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.516 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.516 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.646 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-21T06:57:00.647 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-radosgw.target → /lib/systemd/system/ceph-radosgw.target. 2026-03-21T06:57:00.693 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:00.708 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:00.709 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:00.720 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.720 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:00 vm02 bash[17657]: cluster 2026-03-21T06:56:59.615390+0000 mgr.x (mgr.14152) 401 : cluster [DBG] pgmap v324: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:00.720 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:00 vm02 bash[17657]: cluster 2026-03-21T06:56:59.615390+0000 mgr.x (mgr.14152) 401 : cluster [DBG] pgmap v324: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:00.720 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.720 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.721 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.721 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:57:00 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.721 INFO:teuthology.orchestra.run.vm02.stdout:Setting up ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:00.768 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.769 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.769 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.769 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.769 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:00 vm07 bash[19945]: cluster 2026-03-21T06:56:59.615390+0000 mgr.x (mgr.14152) 401 : cluster [DBG] pgmap v324: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:00.769 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:00 vm07 bash[19945]: cluster 2026-03-21T06:56:59.615390+0000 mgr.x (mgr.14152) 401 : cluster [DBG] pgmap v324: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:00.769 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.769 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:00.835 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T06:57:00.909 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T06:57:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:00 vm04 bash[20194]: cluster 2026-03-21T06:56:59.615390+0000 mgr.x (mgr.14152) 401 : cluster [DBG] pgmap v324: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:00 vm04 bash[20194]: cluster 2026-03-21T06:56:59.615390+0000 mgr.x (mgr.14152) 401 : cluster [DBG] pgmap v324: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:01.032 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.032 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.032 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.032 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.032 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.032 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.032 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.032 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.033 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:00 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.034 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:01.122 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-crash.service → /lib/systemd/system/ceph-crash.service. 2026-03-21T06:57:01.308 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.308 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.308 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.308 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.308 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.501 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:01.501 INFO:teuthology.orchestra.run.vm02.stdout:Running kernel seems to be up-to-date. 2026-03-21T06:57:01.501 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:01.501 INFO:teuthology.orchestra.run.vm02.stdout:Services to be restarted: 2026-03-21T06:57:01.503 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart apache-htcacheclean.service 2026-03-21T06:57:01.508 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart rsyslog.service 2026-03-21T06:57:01.511 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:01.511 INFO:teuthology.orchestra.run.vm02.stdout:Service restarts being deferred: 2026-03-21T06:57:01.511 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart networkd-dispatcher.service 2026-03-21T06:57:01.511 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart unattended-upgrades.service 2026-03-21T06:57:01.511 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:01.511 INFO:teuthology.orchestra.run.vm02.stdout:No containers need to be restarted. 2026-03-21T06:57:01.511 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:01.511 INFO:teuthology.orchestra.run.vm02.stdout:No user sessions are running outdated binaries. 2026-03-21T06:57:01.511 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:01.511 INFO:teuthology.orchestra.run.vm02.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-21T06:57:01.521 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:01.579 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-21T06:57:01.579 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mds.target → /lib/systemd/system/ceph-mds.target. 2026-03-21T06:57:01.580 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.580 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.580 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.580 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.580 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.875 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.875 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.875 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.875 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.875 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:01.964 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:02.034 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-21T06:57:02.034 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mgr.target → /lib/systemd/system/ceph-mgr.target. 2026-03-21T06:57:02.130 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.130 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.131 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.131 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.131 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:01 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.316 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T06:57:02.318 DEBUG:teuthology.orchestra.run.vm02:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install open-iscsi multipath-tools python3-jmespath python3-xmltodict s3cmd 2026-03-21T06:57:02.391 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.391 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.392 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.392 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.392 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.392 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.392 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.392 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.392 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.392 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.393 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T06:57:02.394 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:02.471 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-21T06:57:02.471 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-osd.target → /lib/systemd/system/ceph-osd.target. 2026-03-21T06:57:02.580 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T06:57:02.580 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T06:57:02.651 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.651 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.651 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.651 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.715 INFO:teuthology.orchestra.run.vm02.stdout:open-iscsi is already the newest version (2.1.5-1ubuntu1.1). 2026-03-21T06:57:02.715 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T06:57:02.716 INFO:teuthology.orchestra.run.vm02.stdout: libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T06:57:02.716 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T06:57:02.717 INFO:teuthology.orchestra.run.vm02.stdout:Suggested packages: 2026-03-21T06:57:02.717 INFO:teuthology.orchestra.run.vm02.stdout: multipath-tools-boot 2026-03-21T06:57:02.730 INFO:teuthology.orchestra.run.vm02.stdout:The following NEW packages will be installed: 2026-03-21T06:57:02.730 INFO:teuthology.orchestra.run.vm02.stdout: multipath-tools python3-jmespath python3-xmltodict s3cmd 2026-03-21T06:57:02.755 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 4 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T06:57:02.755 INFO:teuthology.orchestra.run.vm02.stdout:Need to get 486 kB of archives. 2026-03-21T06:57:02.755 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 1930 kB of additional disk space will be used. 2026-03-21T06:57:02.755 INFO:teuthology.orchestra.run.vm02.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-21T06:57:02.770 INFO:teuthology.orchestra.run.vm02.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-21T06:57:02.772 INFO:teuthology.orchestra.run.vm02.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy/universe amd64 s3cmd all 2.2.0-1 [120 kB] 2026-03-21T06:57:02.788 INFO:teuthology.orchestra.run.vm02.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 multipath-tools amd64 0.8.8-1ubuntu1.22.04.4 [331 kB] 2026-03-21T06:57:02.867 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:02.869 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:02.882 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:02.944 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/multi-user.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-21T06:57:02.945 INFO:teuthology.orchestra.run.vm07.stdout:Created symlink /etc/systemd/system/ceph.target.wants/ceph-mon.target → /lib/systemd/system/ceph-mon.target. 2026-03-21T06:57:02.945 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:02 vm07 bash[19945]: cluster 2026-03-21T06:57:01.615687+0000 mgr.x (mgr.14152) 402 : cluster [DBG] pgmap v325: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:02.945 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:02 vm07 bash[19945]: cluster 2026-03-21T06:57:01.615687+0000 mgr.x (mgr.14152) 402 : cluster [DBG] pgmap v325: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:02.945 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:02 vm07 bash[19945]: audit 2026-03-21T06:57:01.769155+0000 mon.c (mon.1) 35 : audit [DBG] from='client.? 192.168.123.107:0/1368721628' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:57:02.945 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:02 vm07 bash[19945]: audit 2026-03-21T06:57:01.769155+0000 mon.c (mon.1) 35 : audit [DBG] from='client.? 192.168.123.107:0/1368721628' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:57:02.945 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.945 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.945 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.945 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.945 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:02 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:02.976 INFO:teuthology.orchestra.run.vm02.stdout:Fetched 486 kB in 0s (7134 kB/s) 2026-03-21T06:57:02.992 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-jmespath. 2026-03-21T06:57:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:02 vm04 bash[20194]: cluster 2026-03-21T06:57:01.615687+0000 mgr.x (mgr.14152) 402 : cluster [DBG] pgmap v325: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:02 vm04 bash[20194]: cluster 2026-03-21T06:57:01.615687+0000 mgr.x (mgr.14152) 402 : cluster [DBG] pgmap v325: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:02 vm04 bash[20194]: audit 2026-03-21T06:57:01.769155+0000 mon.c (mon.1) 35 : audit [DBG] from='client.? 192.168.123.107:0/1368721628' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:57:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:02 vm04 bash[20194]: audit 2026-03-21T06:57:01.769155+0000 mon.c (mon.1) 35 : audit [DBG] from='client.? 192.168.123.107:0/1368721628' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:57:03.023 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126082 files and directories currently installed.) 2026-03-21T06:57:03.025 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-21T06:57:03.026 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-21T06:57:03.043 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-21T06:57:03.049 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-21T06:57:03.050 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-21T06:57:03.064 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package s3cmd. 2026-03-21T06:57:03.070 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../archives/s3cmd_2.2.0-1_all.deb ... 2026-03-21T06:57:03.071 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking s3cmd (2.2.0-1) ... 2026-03-21T06:57:03.095 INFO:teuthology.orchestra.run.vm02.stdout:Selecting previously unselected package multipath-tools. 2026-03-21T06:57:03.101 INFO:teuthology.orchestra.run.vm02.stdout:Preparing to unpack .../multipath-tools_0.8.8-1ubuntu1.22.04.4_amd64.deb ... 2026-03-21T06:57:03.107 INFO:teuthology.orchestra.run.vm02.stdout:Unpacking multipath-tools (0.8.8-1ubuntu1.22.04.4) ... 2026-03-21T06:57:03.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:02 vm02 bash[17657]: cluster 2026-03-21T06:57:01.615687+0000 mgr.x (mgr.14152) 402 : cluster [DBG] pgmap v325: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:03.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:02 vm02 bash[17657]: cluster 2026-03-21T06:57:01.615687+0000 mgr.x (mgr.14152) 402 : cluster [DBG] pgmap v325: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:03.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:02 vm02 bash[17657]: audit 2026-03-21T06:57:01.769155+0000 mon.c (mon.1) 35 : audit [DBG] from='client.? 192.168.123.107:0/1368721628' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:57:03.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:02 vm02 bash[17657]: audit 2026-03-21T06:57:01.769155+0000 mon.c (mon.1) 35 : audit [DBG] from='client.? 192.168.123.107:0/1368721628' entity='client.admin' cmd={"prefix": "status"} : dispatch 2026-03-21T06:57:03.149 INFO:teuthology.orchestra.run.vm02.stdout:Setting up s3cmd (2.2.0-1) ... 2026-03-21T06:57:03.221 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:03 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.221 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:03 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.221 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:03 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.221 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:03 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.221 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:03 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.239 INFO:teuthology.orchestra.run.vm02.stdout:Setting up multipath-tools (0.8.8-1ubuntu1.22.04.4) ... 2026-03-21T06:57:03.316 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:03.329 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:03.331 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:03.346 INFO:teuthology.orchestra.run.vm07.stdout:Setting up ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T06:57:03.468 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T06:57:03.544 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T06:57:03.593 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:57:03 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.593 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:57:03 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.593 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:57:03 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.594 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:57:03 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.594 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:03 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:03 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.651 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:03 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.651 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:03 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.651 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:03 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.651 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:03 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.710 INFO:teuthology.orchestra.run.vm02.stdout:Could not execute systemctl: at /usr/bin/deb-systemd-invoke line 142. 2026-03-21T06:57:03.714 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-21T06:57:03.781 INFO:teuthology.orchestra.run.vm02.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-21T06:57:03.851 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T06:57:03.888 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 06:57:03 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.888 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 06:57:03 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.888 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 06:57:03 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.889 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:57:03 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.889 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:03 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:03.919 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T06:57:04.059 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:04.059 INFO:teuthology.orchestra.run.vm07.stdout:Running kernel seems to be up-to-date. 2026-03-21T06:57:04.059 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:04.059 INFO:teuthology.orchestra.run.vm07.stdout:Services to be restarted: 2026-03-21T06:57:04.062 INFO:teuthology.orchestra.run.vm07.stdout: systemctl restart apache-htcacheclean.service 2026-03-21T06:57:04.068 INFO:teuthology.orchestra.run.vm07.stdout: systemctl restart rsyslog.service 2026-03-21T06:57:04.071 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:04.071 INFO:teuthology.orchestra.run.vm07.stdout:Service restarts being deferred: 2026-03-21T06:57:04.071 INFO:teuthology.orchestra.run.vm07.stdout: systemctl restart networkd-dispatcher.service 2026-03-21T06:57:04.071 INFO:teuthology.orchestra.run.vm07.stdout: systemctl restart unattended-upgrades.service 2026-03-21T06:57:04.071 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:04.071 INFO:teuthology.orchestra.run.vm07.stdout:No containers need to be restarted. 2026-03-21T06:57:04.071 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:04.071 INFO:teuthology.orchestra.run.vm07.stdout:No user sessions are running outdated binaries. 2026-03-21T06:57:04.071 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:04.071 INFO:teuthology.orchestra.run.vm07.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-21T06:57:04.246 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:04.246 INFO:teuthology.orchestra.run.vm02.stdout:Running kernel seems to be up-to-date. 2026-03-21T06:57:04.246 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:04.246 INFO:teuthology.orchestra.run.vm02.stdout:Services to be restarted: 2026-03-21T06:57:04.249 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart apache-htcacheclean.service 2026-03-21T06:57:04.255 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart rsyslog.service 2026-03-21T06:57:04.257 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:04.258 INFO:teuthology.orchestra.run.vm02.stdout:Service restarts being deferred: 2026-03-21T06:57:04.258 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart networkd-dispatcher.service 2026-03-21T06:57:04.258 INFO:teuthology.orchestra.run.vm02.stdout: systemctl restart unattended-upgrades.service 2026-03-21T06:57:04.258 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:04.258 INFO:teuthology.orchestra.run.vm02.stdout:No containers need to be restarted. 2026-03-21T06:57:04.258 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:04.258 INFO:teuthology.orchestra.run.vm02.stdout:No user sessions are running outdated binaries. 2026-03-21T06:57:04.258 INFO:teuthology.orchestra.run.vm02.stdout: 2026-03-21T06:57:04.258 INFO:teuthology.orchestra.run.vm02.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-21T06:57:04.955 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T06:57:04.958 DEBUG:teuthology.orchestra.run.vm07:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" install open-iscsi multipath-tools python3-jmespath python3-xmltodict s3cmd 2026-03-21T06:57:05.004 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:04 vm07 bash[19945]: cluster 2026-03-21T06:57:03.616021+0000 mgr.x (mgr.14152) 403 : cluster [DBG] pgmap v326: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:05.004 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:04 vm07 bash[19945]: cluster 2026-03-21T06:57:03.616021+0000 mgr.x (mgr.14152) 403 : cluster [DBG] pgmap v326: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:05.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:04 vm04 bash[20194]: cluster 2026-03-21T06:57:03.616021+0000 mgr.x (mgr.14152) 403 : cluster [DBG] pgmap v326: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:05.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:04 vm04 bash[20194]: cluster 2026-03-21T06:57:03.616021+0000 mgr.x (mgr.14152) 403 : cluster [DBG] pgmap v326: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:05.033 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T06:57:05.106 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T06:57:05.107 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:04 vm02 bash[17657]: cluster 2026-03-21T06:57:03.616021+0000 mgr.x (mgr.14152) 403 : cluster [DBG] pgmap v326: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:05.107 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:04 vm02 bash[17657]: cluster 2026-03-21T06:57:03.616021+0000 mgr.x (mgr.14152) 403 : cluster [DBG] pgmap v326: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:05.109 DEBUG:teuthology.parallel:result is None 2026-03-21T06:57:05.197 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T06:57:05.198 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T06:57:05.315 INFO:teuthology.orchestra.run.vm07.stdout:open-iscsi is already the newest version (2.1.5-1ubuntu1.1). 2026-03-21T06:57:05.315 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T06:57:05.315 INFO:teuthology.orchestra.run.vm07.stdout: libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T06:57:05.315 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T06:57:05.316 INFO:teuthology.orchestra.run.vm07.stdout:Suggested packages: 2026-03-21T06:57:05.316 INFO:teuthology.orchestra.run.vm07.stdout: multipath-tools-boot 2026-03-21T06:57:05.330 INFO:teuthology.orchestra.run.vm07.stdout:The following NEW packages will be installed: 2026-03-21T06:57:05.330 INFO:teuthology.orchestra.run.vm07.stdout: multipath-tools python3-jmespath python3-xmltodict s3cmd 2026-03-21T06:57:05.354 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 4 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T06:57:05.354 INFO:teuthology.orchestra.run.vm07.stdout:Need to get 486 kB of archives. 2026-03-21T06:57:05.354 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 1930 kB of additional disk space will be used. 2026-03-21T06:57:05.354 INFO:teuthology.orchestra.run.vm07.stdout:Get:1 http://archive.ubuntu.com/ubuntu jammy/main amd64 python3-jmespath all 0.10.0-1 [21.7 kB] 2026-03-21T06:57:05.370 INFO:teuthology.orchestra.run.vm07.stdout:Get:2 http://archive.ubuntu.com/ubuntu jammy/universe amd64 python3-xmltodict all 0.12.0-2 [12.6 kB] 2026-03-21T06:57:05.371 INFO:teuthology.orchestra.run.vm07.stdout:Get:3 http://archive.ubuntu.com/ubuntu jammy/universe amd64 s3cmd all 2.2.0-1 [120 kB] 2026-03-21T06:57:05.388 INFO:teuthology.orchestra.run.vm07.stdout:Get:4 http://archive.ubuntu.com/ubuntu jammy-updates/main amd64 multipath-tools amd64 0.8.8-1ubuntu1.22.04.4 [331 kB] 2026-03-21T06:57:05.573 INFO:teuthology.orchestra.run.vm07.stdout:Fetched 486 kB in 0s (7083 kB/s) 2026-03-21T06:57:05.586 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-jmespath. 2026-03-21T06:57:05.615 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126082 files and directories currently installed.) 2026-03-21T06:57:05.617 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../python3-jmespath_0.10.0-1_all.deb ... 2026-03-21T06:57:05.618 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-jmespath (0.10.0-1) ... 2026-03-21T06:57:05.633 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package python3-xmltodict. 2026-03-21T06:57:05.639 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../python3-xmltodict_0.12.0-2_all.deb ... 2026-03-21T06:57:05.639 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking python3-xmltodict (0.12.0-2) ... 2026-03-21T06:57:05.654 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package s3cmd. 2026-03-21T06:57:05.661 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../archives/s3cmd_2.2.0-1_all.deb ... 2026-03-21T06:57:05.661 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking s3cmd (2.2.0-1) ... 2026-03-21T06:57:05.685 INFO:teuthology.orchestra.run.vm07.stdout:Selecting previously unselected package multipath-tools. 2026-03-21T06:57:05.691 INFO:teuthology.orchestra.run.vm07.stdout:Preparing to unpack .../multipath-tools_0.8.8-1ubuntu1.22.04.4_amd64.deb ... 2026-03-21T06:57:05.696 INFO:teuthology.orchestra.run.vm07.stdout:Unpacking multipath-tools (0.8.8-1ubuntu1.22.04.4) ... 2026-03-21T06:57:05.740 INFO:teuthology.orchestra.run.vm07.stdout:Setting up s3cmd (2.2.0-1) ... 2026-03-21T06:57:05.824 INFO:teuthology.orchestra.run.vm07.stdout:Setting up multipath-tools (0.8.8-1ubuntu1.22.04.4) ... 2026-03-21T06:57:06.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:05 vm04 bash[20194]: cluster 2026-03-21T06:57:05.616374+0000 mgr.x (mgr.14152) 404 : cluster [DBG] pgmap v327: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:06.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:05 vm04 bash[20194]: cluster 2026-03-21T06:57:05.616374+0000 mgr.x (mgr.14152) 404 : cluster [DBG] pgmap v327: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:06.078 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:05 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:06.078 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:05 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:06.078 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:05 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:06.079 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:05 vm07 bash[19945]: cluster 2026-03-21T06:57:05.616374+0000 mgr.x (mgr.14152) 404 : cluster [DBG] pgmap v327: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:06.079 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:05 vm07 bash[19945]: cluster 2026-03-21T06:57:05.616374+0000 mgr.x (mgr.14152) 404 : cluster [DBG] pgmap v327: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:06.079 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:05 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:06.079 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:05 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:06.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:05 vm02 bash[17657]: cluster 2026-03-21T06:57:05.616374+0000 mgr.x (mgr.14152) 404 : cluster [DBG] pgmap v327: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:06.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:05 vm02 bash[17657]: cluster 2026-03-21T06:57:05.616374+0000 mgr.x (mgr.14152) 404 : cluster [DBG] pgmap v327: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:06.315 INFO:teuthology.orchestra.run.vm07.stdout:Could not execute systemctl: at /usr/bin/deb-systemd-invoke line 142. 2026-03-21T06:57:06.321 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-xmltodict (0.12.0-2) ... 2026-03-21T06:57:06.387 INFO:teuthology.orchestra.run.vm07.stdout:Setting up python3-jmespath (0.10.0-1) ... 2026-03-21T06:57:06.401 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 06:57:06 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:06.401 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 06:57:06 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:06.401 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:06 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:06.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:06 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:06.401 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 06:57:06 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T06:57:06.454 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T06:57:06.522 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T06:57:06.884 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:06.884 INFO:teuthology.orchestra.run.vm07.stdout:Running kernel seems to be up-to-date. 2026-03-21T06:57:06.884 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:06.884 INFO:teuthology.orchestra.run.vm07.stdout:Services to be restarted: 2026-03-21T06:57:06.887 INFO:teuthology.orchestra.run.vm07.stdout: systemctl restart apache-htcacheclean.service 2026-03-21T06:57:06.894 INFO:teuthology.orchestra.run.vm07.stdout: systemctl restart rsyslog.service 2026-03-21T06:57:06.897 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:06.897 INFO:teuthology.orchestra.run.vm07.stdout:Service restarts being deferred: 2026-03-21T06:57:06.897 INFO:teuthology.orchestra.run.vm07.stdout: systemctl restart networkd-dispatcher.service 2026-03-21T06:57:06.897 INFO:teuthology.orchestra.run.vm07.stdout: systemctl restart unattended-upgrades.service 2026-03-21T06:57:06.897 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:06.897 INFO:teuthology.orchestra.run.vm07.stdout:No containers need to be restarted. 2026-03-21T06:57:06.897 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:06.897 INFO:teuthology.orchestra.run.vm07.stdout:No user sessions are running outdated binaries. 2026-03-21T06:57:06.897 INFO:teuthology.orchestra.run.vm07.stdout: 2026-03-21T06:57:06.897 INFO:teuthology.orchestra.run.vm07.stdout:No VM guests are running outdated hypervisor (qemu) binaries on this host. 2026-03-21T06:57:07.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:57:06 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:57:07.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:06 vm02 bash[17657]: audit 2026-03-21T06:57:06.678141+0000 mgr.x (mgr.14152) 405 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:07.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:06 vm02 bash[17657]: audit 2026-03-21T06:57:06.678141+0000 mgr.x (mgr.14152) 405 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:06 vm07 bash[19945]: audit 2026-03-21T06:57:06.678141+0000 mgr.x (mgr.14152) 405 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:06 vm07 bash[19945]: audit 2026-03-21T06:57:06.678141+0000 mgr.x (mgr.14152) 405 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:07.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:06 vm04 bash[20194]: audit 2026-03-21T06:57:06.678141+0000 mgr.x (mgr.14152) 405 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:07.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:06 vm04 bash[20194]: audit 2026-03-21T06:57:06.678141+0000 mgr.x (mgr.14152) 405 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:07.645 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T06:57:07.649 DEBUG:teuthology.parallel:result is None 2026-03-21T06:57:07.649 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-21T06:57:07.867 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:07 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:57:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:07 vm07 bash[19945]: audit 2026-03-21T06:57:07.510130+0000 mgr.x (mgr.14152) 406 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:07 vm07 bash[19945]: audit 2026-03-21T06:57:07.510130+0000 mgr.x (mgr.14152) 406 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:07 vm07 bash[19945]: cluster 2026-03-21T06:57:07.616727+0000 mgr.x (mgr.14152) 407 : cluster [DBG] pgmap v328: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:07 vm07 bash[19945]: cluster 2026-03-21T06:57:07.616727+0000 mgr.x (mgr.14152) 407 : cluster [DBG] pgmap v328: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:08.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:07 vm02 bash[17657]: audit 2026-03-21T06:57:07.510130+0000 mgr.x (mgr.14152) 406 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:08.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:07 vm02 bash[17657]: audit 2026-03-21T06:57:07.510130+0000 mgr.x (mgr.14152) 406 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:08.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:07 vm02 bash[17657]: cluster 2026-03-21T06:57:07.616727+0000 mgr.x (mgr.14152) 407 : cluster [DBG] pgmap v328: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:08.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:07 vm02 bash[17657]: cluster 2026-03-21T06:57:07.616727+0000 mgr.x (mgr.14152) 407 : cluster [DBG] pgmap v328: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:08.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:07 vm04 bash[20194]: audit 2026-03-21T06:57:07.510130+0000 mgr.x (mgr.14152) 406 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:08.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:07 vm04 bash[20194]: audit 2026-03-21T06:57:07.510130+0000 mgr.x (mgr.14152) 406 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:08.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:07 vm04 bash[20194]: cluster 2026-03-21T06:57:07.616727+0000 mgr.x (mgr.14152) 407 : cluster [DBG] pgmap v328: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:08.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:07 vm04 bash[20194]: cluster 2026-03-21T06:57:07.616727+0000 mgr.x (mgr.14152) 407 : cluster [DBG] pgmap v328: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:08.304 DEBUG:teuthology.orchestra.run.vm02:> dpkg-query -W -f '${Version}' ceph 2026-03-21T06:57:08.312 INFO:teuthology.orchestra.run.vm02.stdout:20.2.0-712-g70f8415b-1jammy 2026-03-21T06:57:08.312 INFO:teuthology.packaging:The installed version of ceph is 20.2.0-712-g70f8415b-1jammy 2026-03-21T06:57:08.312 INFO:teuthology.task.install:The correct ceph version 20.2.0-712-g70f8415b-1jammy is installed. 2026-03-21T06:57:08.313 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-21T06:57:08.956 DEBUG:teuthology.orchestra.run.vm04:> dpkg-query -W -f '${Version}' ceph 2026-03-21T06:57:08.965 INFO:teuthology.orchestra.run.vm04.stdout:20.2.0-712-g70f8415b-1jammy 2026-03-21T06:57:08.965 INFO:teuthology.packaging:The installed version of ceph is 20.2.0-712-g70f8415b-1jammy 2026-03-21T06:57:08.965 INFO:teuthology.task.install:The correct ceph version 20.2.0-712-g70f8415b-1jammy is installed. 2026-03-21T06:57:08.966 DEBUG:teuthology.packaging:Querying https://shaman.ceph.com/api/search?status=ready&project=ceph&flavor=default&distros=ubuntu%2F22.04%2Fx86_64&sha1=70f8415b300f041766fa27faf7d5472699e32388 2026-03-21T06:57:09.637 DEBUG:teuthology.orchestra.run.vm07:> dpkg-query -W -f '${Version}' ceph 2026-03-21T06:57:09.648 INFO:teuthology.orchestra.run.vm07.stdout:20.2.0-712-g70f8415b-1jammy 2026-03-21T06:57:09.648 INFO:teuthology.packaging:The installed version of ceph is 20.2.0-712-g70f8415b-1jammy 2026-03-21T06:57:09.649 INFO:teuthology.task.install:The correct ceph version 20.2.0-712-g70f8415b-1jammy is installed. 2026-03-21T06:57:09.650 INFO:teuthology.task.install.util:Shipping valgrind.supp... 2026-03-21T06:57:09.650 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:57:09.650 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-21T06:57:09.658 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:57:09.659 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-21T06:57:09.666 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:57:09.666 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/home/ubuntu/cephtest/valgrind.supp 2026-03-21T06:57:09.697 INFO:teuthology.task.install.util:Shipping 'daemon-helper'... 2026-03-21T06:57:09.698 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:57:09.698 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/daemon-helper 2026-03-21T06:57:09.707 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-21T06:57:09.756 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:57:09.756 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/daemon-helper 2026-03-21T06:57:09.764 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-21T06:57:09.812 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:57:09.813 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/usr/bin/daemon-helper 2026-03-21T06:57:09.819 DEBUG:teuthology.orchestra.run.vm07:> sudo chmod a=rx -- /usr/bin/daemon-helper 2026-03-21T06:57:09.868 INFO:teuthology.task.install.util:Shipping 'adjust-ulimits'... 2026-03-21T06:57:09.869 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:57:09.869 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-21T06:57:09.877 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-21T06:57:09.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:09 vm02 bash[17657]: audit 2026-03-21T06:57:09.603087+0000 mon.a (mon.0) 854 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:57:09.926 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:09 vm02 bash[17657]: audit 2026-03-21T06:57:09.603087+0000 mon.a (mon.0) 854 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:57:09.926 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:57:09.927 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-21T06:57:09.932 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:09 vm04 bash[20194]: audit 2026-03-21T06:57:09.603087+0000 mon.a (mon.0) 854 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:57:09.933 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:09 vm04 bash[20194]: audit 2026-03-21T06:57:09.603087+0000 mon.a (mon.0) 854 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:57:09.934 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-21T06:57:09.989 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:57:09.989 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/usr/bin/adjust-ulimits 2026-03-21T06:57:09.995 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:09 vm07 bash[19945]: audit 2026-03-21T06:57:09.603087+0000 mon.a (mon.0) 854 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:57:09.995 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:09 vm07 bash[19945]: audit 2026-03-21T06:57:09.603087+0000 mon.a (mon.0) 854 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:57:09.997 DEBUG:teuthology.orchestra.run.vm07:> sudo chmod a=rx -- /usr/bin/adjust-ulimits 2026-03-21T06:57:10.049 INFO:teuthology.task.install.util:Shipping 'stdin-killer'... 2026-03-21T06:57:10.049 DEBUG:teuthology.orchestra.run.vm02:> set -ex 2026-03-21T06:57:10.049 DEBUG:teuthology.orchestra.run.vm02:> sudo dd of=/usr/bin/stdin-killer 2026-03-21T06:57:10.057 DEBUG:teuthology.orchestra.run.vm02:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-21T06:57:10.107 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:57:10.107 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/usr/bin/stdin-killer 2026-03-21T06:57:10.116 DEBUG:teuthology.orchestra.run.vm04:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-21T06:57:10.164 DEBUG:teuthology.orchestra.run.vm07:> set -ex 2026-03-21T06:57:10.164 DEBUG:teuthology.orchestra.run.vm07:> sudo dd of=/usr/bin/stdin-killer 2026-03-21T06:57:10.172 DEBUG:teuthology.orchestra.run.vm07:> sudo chmod a=rx -- /usr/bin/stdin-killer 2026-03-21T06:57:10.221 INFO:teuthology.run_tasks:Running task ceph_iscsi_client... 2026-03-21T06:57:10.224 INFO:tasks.ceph_iscsi_client:Setting up ceph-iscsi client... 2026-03-21T06:57:10.224 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:57:10.224 DEBUG:teuthology.orchestra.run.vm04:> sudo mkdir -p /etc/iscsi 2026-03-21T06:57:10.224 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/iscsi/initiatorname.iscsi 2026-03-21T06:57:10.237 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl restart iscsid 2026-03-21T06:57:10.295 DEBUG:teuthology.orchestra.run.vm04:> sudo modprobe dm_multipath 2026-03-21T06:57:10.345 DEBUG:teuthology.orchestra.run.vm04:> set -ex 2026-03-21T06:57:10.345 DEBUG:teuthology.orchestra.run.vm04:> sudo dd of=/etc/multipath.conf 2026-03-21T06:57:10.396 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl start multipathd 2026-03-21T06:57:10.445 INFO:teuthology.run_tasks:Running task cram... 2026-03-21T06:57:10.448 INFO:tasks.cram:Pulling tests from https://github.com/kshtsk/ceph.git ref 0392f78529848ec72469e8e431875cb98d3a5fb4 2026-03-21T06:57:10.448 DEBUG:teuthology.orchestra.run.vm02:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.0 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-21T06:57:11.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:10 vm04 bash[20194]: cluster 2026-03-21T06:57:09.617073+0000 mgr.x (mgr.14152) 408 : cluster [DBG] pgmap v329: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:11.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:10 vm04 bash[20194]: cluster 2026-03-21T06:57:09.617073+0000 mgr.x (mgr.14152) 408 : cluster [DBG] pgmap v329: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:11.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:10 vm02 bash[17657]: cluster 2026-03-21T06:57:09.617073+0000 mgr.x (mgr.14152) 408 : cluster [DBG] pgmap v329: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:11.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:10 vm02 bash[17657]: cluster 2026-03-21T06:57:09.617073+0000 mgr.x (mgr.14152) 408 : cluster [DBG] pgmap v329: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:10 vm07 bash[19945]: cluster 2026-03-21T06:57:09.617073+0000 mgr.x (mgr.14152) 408 : cluster [DBG] pgmap v329: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:11.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:10 vm07 bash[19945]: cluster 2026-03-21T06:57:09.617073+0000 mgr.x (mgr.14152) 408 : cluster [DBG] pgmap v329: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:12.093 INFO:teuthology.orchestra.run.vm02.stdout:Collecting cram==0.6 2026-03-21T06:57:12.132 INFO:teuthology.orchestra.run.vm02.stdout: Downloading cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-21T06:57:12.146 INFO:teuthology.orchestra.run.vm02.stdout:Installing collected packages: cram 2026-03-21T06:57:12.152 INFO:teuthology.orchestra.run.vm02.stdout:Successfully installed cram-0.6 2026-03-21T06:57:12.181 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/clone.client.0 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.0 && cd /home/ubuntu/cephtest/clone.client.0 && git checkout 0392f78529848ec72469e8e431875cb98d3a5fb4 2026-03-21T06:57:12.185 INFO:teuthology.orchestra.run.vm02.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.0'... 2026-03-21T06:57:13.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:12 vm04 bash[20194]: cluster 2026-03-21T06:57:11.617443+0000 mgr.x (mgr.14152) 409 : cluster [DBG] pgmap v330: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:13.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:12 vm04 bash[20194]: cluster 2026-03-21T06:57:11.617443+0000 mgr.x (mgr.14152) 409 : cluster [DBG] pgmap v330: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:13.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:12 vm02 bash[17657]: cluster 2026-03-21T06:57:11.617443+0000 mgr.x (mgr.14152) 409 : cluster [DBG] pgmap v330: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:13.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:12 vm02 bash[17657]: cluster 2026-03-21T06:57:11.617443+0000 mgr.x (mgr.14152) 409 : cluster [DBG] pgmap v330: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:12 vm07 bash[19945]: cluster 2026-03-21T06:57:11.617443+0000 mgr.x (mgr.14152) 409 : cluster [DBG] pgmap v330: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:12 vm07 bash[19945]: cluster 2026-03-21T06:57:11.617443+0000 mgr.x (mgr.14152) 409 : cluster [DBG] pgmap v330: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:15.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:14 vm04 bash[20194]: cluster 2026-03-21T06:57:13.617769+0000 mgr.x (mgr.14152) 410 : cluster [DBG] pgmap v331: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:15.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:14 vm04 bash[20194]: cluster 2026-03-21T06:57:13.617769+0000 mgr.x (mgr.14152) 410 : cluster [DBG] pgmap v331: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:15.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:14 vm04 bash[20194]: audit 2026-03-21T06:57:14.649706+0000 mon.a (mon.0) 855 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:15.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:14 vm04 bash[20194]: audit 2026-03-21T06:57:14.649706+0000 mon.a (mon.0) 855 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:15.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:14 vm04 bash[20194]: audit 2026-03-21T06:57:14.655002+0000 mon.a (mon.0) 856 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:15.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:14 vm04 bash[20194]: audit 2026-03-21T06:57:14.655002+0000 mon.a (mon.0) 856 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:15.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:14 vm02 bash[17657]: cluster 2026-03-21T06:57:13.617769+0000 mgr.x (mgr.14152) 410 : cluster [DBG] pgmap v331: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:15.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:14 vm02 bash[17657]: cluster 2026-03-21T06:57:13.617769+0000 mgr.x (mgr.14152) 410 : cluster [DBG] pgmap v331: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:15.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:14 vm02 bash[17657]: audit 2026-03-21T06:57:14.649706+0000 mon.a (mon.0) 855 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:15.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:14 vm02 bash[17657]: audit 2026-03-21T06:57:14.649706+0000 mon.a (mon.0) 855 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:15.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:14 vm02 bash[17657]: audit 2026-03-21T06:57:14.655002+0000 mon.a (mon.0) 856 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:15.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:14 vm02 bash[17657]: audit 2026-03-21T06:57:14.655002+0000 mon.a (mon.0) 856 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:14 vm07 bash[19945]: cluster 2026-03-21T06:57:13.617769+0000 mgr.x (mgr.14152) 410 : cluster [DBG] pgmap v331: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:14 vm07 bash[19945]: cluster 2026-03-21T06:57:13.617769+0000 mgr.x (mgr.14152) 410 : cluster [DBG] pgmap v331: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:14 vm07 bash[19945]: audit 2026-03-21T06:57:14.649706+0000 mon.a (mon.0) 855 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:14 vm07 bash[19945]: audit 2026-03-21T06:57:14.649706+0000 mon.a (mon.0) 855 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:14 vm07 bash[19945]: audit 2026-03-21T06:57:14.655002+0000 mon.a (mon.0) 856 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:14 vm07 bash[19945]: audit 2026-03-21T06:57:14.655002+0000 mon.a (mon.0) 856 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: cluster 2026-03-21T06:57:15.618123+0000 mgr.x (mgr.14152) 411 : cluster [DBG] pgmap v332: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: cluster 2026-03-21T06:57:15.618123+0000 mgr.x (mgr.14152) 411 : cluster [DBG] pgmap v332: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.265747+0000 mon.a (mon.0) 857 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.265747+0000 mon.a (mon.0) 857 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.270906+0000 mon.a (mon.0) 858 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.270906+0000 mon.a (mon.0) 858 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.271902+0000 mon.a (mon.0) 859 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.271902+0000 mon.a (mon.0) 859 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.272597+0000 mon.a (mon.0) 860 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.272597+0000 mon.a (mon.0) 860 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.276233+0000 mon.a (mon.0) 861 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.276233+0000 mon.a (mon.0) 861 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.277199+0000 mon.a (mon.0) 862 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:57:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:16 vm04 bash[20194]: audit 2026-03-21T06:57:16.277199+0000 mon.a (mon.0) 862 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: cluster 2026-03-21T06:57:15.618123+0000 mgr.x (mgr.14152) 411 : cluster [DBG] pgmap v332: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: cluster 2026-03-21T06:57:15.618123+0000 mgr.x (mgr.14152) 411 : cluster [DBG] pgmap v332: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.265747+0000 mon.a (mon.0) 857 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.265747+0000 mon.a (mon.0) 857 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.270906+0000 mon.a (mon.0) 858 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.270906+0000 mon.a (mon.0) 858 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.271902+0000 mon.a (mon.0) 859 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.271902+0000 mon.a (mon.0) 859 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.272597+0000 mon.a (mon.0) 860 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.272597+0000 mon.a (mon.0) 860 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.276233+0000 mon.a (mon.0) 861 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.276233+0000 mon.a (mon.0) 861 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.277199+0000 mon.a (mon.0) 862 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:57:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[17657]: audit 2026-03-21T06:57:16.277199+0000 mon.a (mon.0) 862 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:57:17.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:57:16 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: cluster 2026-03-21T06:57:15.618123+0000 mgr.x (mgr.14152) 411 : cluster [DBG] pgmap v332: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: cluster 2026-03-21T06:57:15.618123+0000 mgr.x (mgr.14152) 411 : cluster [DBG] pgmap v332: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.265747+0000 mon.a (mon.0) 857 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.265747+0000 mon.a (mon.0) 857 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.270906+0000 mon.a (mon.0) 858 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.270906+0000 mon.a (mon.0) 858 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.271902+0000 mon.a (mon.0) 859 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.271902+0000 mon.a (mon.0) 859 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.272597+0000 mon.a (mon.0) 860 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.272597+0000 mon.a (mon.0) 860 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.276233+0000 mon.a (mon.0) 861 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.276233+0000 mon.a (mon.0) 861 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.277199+0000 mon.a (mon.0) 862 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:57:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:16 vm07 bash[19945]: audit 2026-03-21T06:57:16.277199+0000 mon.a (mon.0) 862 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:57:17.901 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:17 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:57:17.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:17 vm07 bash[19945]: audit 2026-03-21T06:57:16.688631+0000 mgr.x (mgr.14152) 412 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:17.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:17 vm07 bash[19945]: audit 2026-03-21T06:57:16.688631+0000 mgr.x (mgr.14152) 412 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:18.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:17 vm04 bash[20194]: audit 2026-03-21T06:57:16.688631+0000 mgr.x (mgr.14152) 412 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:18.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:17 vm04 bash[20194]: audit 2026-03-21T06:57:16.688631+0000 mgr.x (mgr.14152) 412 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:18.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:17 vm02 bash[17657]: audit 2026-03-21T06:57:16.688631+0000 mgr.x (mgr.14152) 412 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:18.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:17 vm02 bash[17657]: audit 2026-03-21T06:57:16.688631+0000 mgr.x (mgr.14152) 412 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:19.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:18 vm04 bash[20194]: audit 2026-03-21T06:57:17.521056+0000 mgr.x (mgr.14152) 413 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:19.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:18 vm04 bash[20194]: audit 2026-03-21T06:57:17.521056+0000 mgr.x (mgr.14152) 413 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:19.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:18 vm04 bash[20194]: cluster 2026-03-21T06:57:17.618431+0000 mgr.x (mgr.14152) 414 : cluster [DBG] pgmap v333: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:19.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:18 vm04 bash[20194]: cluster 2026-03-21T06:57:17.618431+0000 mgr.x (mgr.14152) 414 : cluster [DBG] pgmap v333: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:19.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:18 vm02 bash[17657]: audit 2026-03-21T06:57:17.521056+0000 mgr.x (mgr.14152) 413 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:19.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:18 vm02 bash[17657]: audit 2026-03-21T06:57:17.521056+0000 mgr.x (mgr.14152) 413 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:19.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:18 vm02 bash[17657]: cluster 2026-03-21T06:57:17.618431+0000 mgr.x (mgr.14152) 414 : cluster [DBG] pgmap v333: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:19.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:18 vm02 bash[17657]: cluster 2026-03-21T06:57:17.618431+0000 mgr.x (mgr.14152) 414 : cluster [DBG] pgmap v333: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:18 vm07 bash[19945]: audit 2026-03-21T06:57:17.521056+0000 mgr.x (mgr.14152) 413 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:18 vm07 bash[19945]: audit 2026-03-21T06:57:17.521056+0000 mgr.x (mgr.14152) 413 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:18 vm07 bash[19945]: cluster 2026-03-21T06:57:17.618431+0000 mgr.x (mgr.14152) 414 : cluster [DBG] pgmap v333: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:18 vm07 bash[19945]: cluster 2026-03-21T06:57:17.618431+0000 mgr.x (mgr.14152) 414 : cluster [DBG] pgmap v333: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:20.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:19 vm04 bash[20194]: cluster 2026-03-21T06:57:19.618951+0000 mgr.x (mgr.14152) 415 : cluster [DBG] pgmap v334: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:20.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:19 vm04 bash[20194]: cluster 2026-03-21T06:57:19.618951+0000 mgr.x (mgr.14152) 415 : cluster [DBG] pgmap v334: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:20.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:19 vm02 bash[17657]: cluster 2026-03-21T06:57:19.618951+0000 mgr.x (mgr.14152) 415 : cluster [DBG] pgmap v334: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:20.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:19 vm02 bash[17657]: cluster 2026-03-21T06:57:19.618951+0000 mgr.x (mgr.14152) 415 : cluster [DBG] pgmap v334: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:20.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:19 vm07 bash[19945]: cluster 2026-03-21T06:57:19.618951+0000 mgr.x (mgr.14152) 415 : cluster [DBG] pgmap v334: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:20.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:19 vm07 bash[19945]: cluster 2026-03-21T06:57:19.618951+0000 mgr.x (mgr.14152) 415 : cluster [DBG] pgmap v334: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:23.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:22 vm02 bash[17657]: cluster 2026-03-21T06:57:21.619301+0000 mgr.x (mgr.14152) 416 : cluster [DBG] pgmap v335: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:23.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:22 vm02 bash[17657]: cluster 2026-03-21T06:57:21.619301+0000 mgr.x (mgr.14152) 416 : cluster [DBG] pgmap v335: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:22 vm07 bash[19945]: cluster 2026-03-21T06:57:21.619301+0000 mgr.x (mgr.14152) 416 : cluster [DBG] pgmap v335: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:23.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:22 vm07 bash[19945]: cluster 2026-03-21T06:57:21.619301+0000 mgr.x (mgr.14152) 416 : cluster [DBG] pgmap v335: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:23.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:22 vm04 bash[20194]: cluster 2026-03-21T06:57:21.619301+0000 mgr.x (mgr.14152) 416 : cluster [DBG] pgmap v335: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:23.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:22 vm04 bash[20194]: cluster 2026-03-21T06:57:21.619301+0000 mgr.x (mgr.14152) 416 : cluster [DBG] pgmap v335: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:24.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:23 vm02 bash[17657]: cluster 2026-03-21T06:57:23.619595+0000 mgr.x (mgr.14152) 417 : cluster [DBG] pgmap v336: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:24.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:23 vm02 bash[17657]: cluster 2026-03-21T06:57:23.619595+0000 mgr.x (mgr.14152) 417 : cluster [DBG] pgmap v336: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:24.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:23 vm07 bash[19945]: cluster 2026-03-21T06:57:23.619595+0000 mgr.x (mgr.14152) 417 : cluster [DBG] pgmap v336: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:24.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:23 vm07 bash[19945]: cluster 2026-03-21T06:57:23.619595+0000 mgr.x (mgr.14152) 417 : cluster [DBG] pgmap v336: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:24.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:23 vm04 bash[20194]: cluster 2026-03-21T06:57:23.619595+0000 mgr.x (mgr.14152) 417 : cluster [DBG] pgmap v336: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:24.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:23 vm04 bash[20194]: cluster 2026-03-21T06:57:23.619595+0000 mgr.x (mgr.14152) 417 : cluster [DBG] pgmap v336: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:27.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:26 vm04 bash[20194]: cluster 2026-03-21T06:57:25.619903+0000 mgr.x (mgr.14152) 418 : cluster [DBG] pgmap v337: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:27.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:26 vm04 bash[20194]: cluster 2026-03-21T06:57:25.619903+0000 mgr.x (mgr.14152) 418 : cluster [DBG] pgmap v337: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:57:26 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:57:27.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:26 vm02 bash[17657]: cluster 2026-03-21T06:57:25.619903+0000 mgr.x (mgr.14152) 418 : cluster [DBG] pgmap v337: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:27.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:26 vm02 bash[17657]: cluster 2026-03-21T06:57:25.619903+0000 mgr.x (mgr.14152) 418 : cluster [DBG] pgmap v337: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:26 vm07 bash[19945]: cluster 2026-03-21T06:57:25.619903+0000 mgr.x (mgr.14152) 418 : cluster [DBG] pgmap v337: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:26 vm07 bash[19945]: cluster 2026-03-21T06:57:25.619903+0000 mgr.x (mgr.14152) 418 : cluster [DBG] pgmap v337: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:27.901 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:27 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:57:27.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:27 vm07 bash[19945]: audit 2026-03-21T06:57:26.695386+0000 mgr.x (mgr.14152) 419 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:27.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:27 vm07 bash[19945]: audit 2026-03-21T06:57:26.695386+0000 mgr.x (mgr.14152) 419 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:28.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:27 vm04 bash[20194]: audit 2026-03-21T06:57:26.695386+0000 mgr.x (mgr.14152) 419 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:28.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:27 vm04 bash[20194]: audit 2026-03-21T06:57:26.695386+0000 mgr.x (mgr.14152) 419 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:28.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:27 vm02 bash[17657]: audit 2026-03-21T06:57:26.695386+0000 mgr.x (mgr.14152) 419 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:28.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:27 vm02 bash[17657]: audit 2026-03-21T06:57:26.695386+0000 mgr.x (mgr.14152) 419 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:29.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:28 vm04 bash[20194]: audit 2026-03-21T06:57:27.524989+0000 mgr.x (mgr.14152) 420 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:29.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:28 vm04 bash[20194]: audit 2026-03-21T06:57:27.524989+0000 mgr.x (mgr.14152) 420 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:29.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:28 vm04 bash[20194]: cluster 2026-03-21T06:57:27.620165+0000 mgr.x (mgr.14152) 421 : cluster [DBG] pgmap v338: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:29.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:28 vm04 bash[20194]: cluster 2026-03-21T06:57:27.620165+0000 mgr.x (mgr.14152) 421 : cluster [DBG] pgmap v338: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:29.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:28 vm02 bash[17657]: audit 2026-03-21T06:57:27.524989+0000 mgr.x (mgr.14152) 420 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:29.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:28 vm02 bash[17657]: audit 2026-03-21T06:57:27.524989+0000 mgr.x (mgr.14152) 420 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:29.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:28 vm02 bash[17657]: cluster 2026-03-21T06:57:27.620165+0000 mgr.x (mgr.14152) 421 : cluster [DBG] pgmap v338: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:29.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:28 vm02 bash[17657]: cluster 2026-03-21T06:57:27.620165+0000 mgr.x (mgr.14152) 421 : cluster [DBG] pgmap v338: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:28 vm07 bash[19945]: audit 2026-03-21T06:57:27.524989+0000 mgr.x (mgr.14152) 420 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:28 vm07 bash[19945]: audit 2026-03-21T06:57:27.524989+0000 mgr.x (mgr.14152) 420 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:28 vm07 bash[19945]: cluster 2026-03-21T06:57:27.620165+0000 mgr.x (mgr.14152) 421 : cluster [DBG] pgmap v338: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:28 vm07 bash[19945]: cluster 2026-03-21T06:57:27.620165+0000 mgr.x (mgr.14152) 421 : cluster [DBG] pgmap v338: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:31.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:30 vm04 bash[20194]: cluster 2026-03-21T06:57:29.620767+0000 mgr.x (mgr.14152) 422 : cluster [DBG] pgmap v339: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:31.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:30 vm04 bash[20194]: cluster 2026-03-21T06:57:29.620767+0000 mgr.x (mgr.14152) 422 : cluster [DBG] pgmap v339: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:31.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:30 vm02 bash[17657]: cluster 2026-03-21T06:57:29.620767+0000 mgr.x (mgr.14152) 422 : cluster [DBG] pgmap v339: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:31.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:30 vm02 bash[17657]: cluster 2026-03-21T06:57:29.620767+0000 mgr.x (mgr.14152) 422 : cluster [DBG] pgmap v339: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:30 vm07 bash[19945]: cluster 2026-03-21T06:57:29.620767+0000 mgr.x (mgr.14152) 422 : cluster [DBG] pgmap v339: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:31.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:30 vm07 bash[19945]: cluster 2026-03-21T06:57:29.620767+0000 mgr.x (mgr.14152) 422 : cluster [DBG] pgmap v339: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:33.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:32 vm04 bash[20194]: cluster 2026-03-21T06:57:31.621106+0000 mgr.x (mgr.14152) 423 : cluster [DBG] pgmap v340: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:33.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:32 vm04 bash[20194]: cluster 2026-03-21T06:57:31.621106+0000 mgr.x (mgr.14152) 423 : cluster [DBG] pgmap v340: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:33.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:32 vm02 bash[17657]: cluster 2026-03-21T06:57:31.621106+0000 mgr.x (mgr.14152) 423 : cluster [DBG] pgmap v340: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:33.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:32 vm02 bash[17657]: cluster 2026-03-21T06:57:31.621106+0000 mgr.x (mgr.14152) 423 : cluster [DBG] pgmap v340: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:32 vm07 bash[19945]: cluster 2026-03-21T06:57:31.621106+0000 mgr.x (mgr.14152) 423 : cluster [DBG] pgmap v340: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:32 vm07 bash[19945]: cluster 2026-03-21T06:57:31.621106+0000 mgr.x (mgr.14152) 423 : cluster [DBG] pgmap v340: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:35.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:34 vm04 bash[20194]: cluster 2026-03-21T06:57:33.621405+0000 mgr.x (mgr.14152) 424 : cluster [DBG] pgmap v341: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:35.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:34 vm04 bash[20194]: cluster 2026-03-21T06:57:33.621405+0000 mgr.x (mgr.14152) 424 : cluster [DBG] pgmap v341: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:35.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:34 vm02 bash[17657]: cluster 2026-03-21T06:57:33.621405+0000 mgr.x (mgr.14152) 424 : cluster [DBG] pgmap v341: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:35.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:34 vm02 bash[17657]: cluster 2026-03-21T06:57:33.621405+0000 mgr.x (mgr.14152) 424 : cluster [DBG] pgmap v341: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:35.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:34 vm07 bash[19945]: cluster 2026-03-21T06:57:33.621405+0000 mgr.x (mgr.14152) 424 : cluster [DBG] pgmap v341: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:35.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:34 vm07 bash[19945]: cluster 2026-03-21T06:57:33.621405+0000 mgr.x (mgr.14152) 424 : cluster [DBG] pgmap v341: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:36 vm04 bash[20194]: cluster 2026-03-21T06:57:35.621788+0000 mgr.x (mgr.14152) 425 : cluster [DBG] pgmap v342: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:37.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:36 vm04 bash[20194]: cluster 2026-03-21T06:57:35.621788+0000 mgr.x (mgr.14152) 425 : cluster [DBG] pgmap v342: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:37.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:36 vm02 bash[17657]: cluster 2026-03-21T06:57:35.621788+0000 mgr.x (mgr.14152) 425 : cluster [DBG] pgmap v342: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:37.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:36 vm02 bash[17657]: cluster 2026-03-21T06:57:35.621788+0000 mgr.x (mgr.14152) 425 : cluster [DBG] pgmap v342: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:37.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:57:36 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:57:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:36 vm07 bash[19945]: cluster 2026-03-21T06:57:35.621788+0000 mgr.x (mgr.14152) 425 : cluster [DBG] pgmap v342: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:36 vm07 bash[19945]: cluster 2026-03-21T06:57:35.621788+0000 mgr.x (mgr.14152) 425 : cluster [DBG] pgmap v342: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:37.901 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:37 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:57:37.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:37 vm07 bash[19945]: audit 2026-03-21T06:57:36.701988+0000 mgr.x (mgr.14152) 426 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:37.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:37 vm07 bash[19945]: audit 2026-03-21T06:57:36.701988+0000 mgr.x (mgr.14152) 426 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:37 vm04 bash[20194]: audit 2026-03-21T06:57:36.701988+0000 mgr.x (mgr.14152) 426 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:38.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:37 vm04 bash[20194]: audit 2026-03-21T06:57:36.701988+0000 mgr.x (mgr.14152) 426 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:38.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:37 vm02 bash[17657]: audit 2026-03-21T06:57:36.701988+0000 mgr.x (mgr.14152) 426 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:38.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:37 vm02 bash[17657]: audit 2026-03-21T06:57:36.701988+0000 mgr.x (mgr.14152) 426 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:38 vm04 bash[20194]: audit 2026-03-21T06:57:37.535824+0000 mgr.x (mgr.14152) 427 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:38 vm04 bash[20194]: audit 2026-03-21T06:57:37.535824+0000 mgr.x (mgr.14152) 427 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:38 vm04 bash[20194]: cluster 2026-03-21T06:57:37.622148+0000 mgr.x (mgr.14152) 428 : cluster [DBG] pgmap v343: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:39.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:38 vm04 bash[20194]: cluster 2026-03-21T06:57:37.622148+0000 mgr.x (mgr.14152) 428 : cluster [DBG] pgmap v343: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:39.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:38 vm02 bash[17657]: audit 2026-03-21T06:57:37.535824+0000 mgr.x (mgr.14152) 427 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:39.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:38 vm02 bash[17657]: audit 2026-03-21T06:57:37.535824+0000 mgr.x (mgr.14152) 427 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:39.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:38 vm02 bash[17657]: cluster 2026-03-21T06:57:37.622148+0000 mgr.x (mgr.14152) 428 : cluster [DBG] pgmap v343: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:39.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:38 vm02 bash[17657]: cluster 2026-03-21T06:57:37.622148+0000 mgr.x (mgr.14152) 428 : cluster [DBG] pgmap v343: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:38 vm07 bash[19945]: audit 2026-03-21T06:57:37.535824+0000 mgr.x (mgr.14152) 427 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:38 vm07 bash[19945]: audit 2026-03-21T06:57:37.535824+0000 mgr.x (mgr.14152) 427 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:38 vm07 bash[19945]: cluster 2026-03-21T06:57:37.622148+0000 mgr.x (mgr.14152) 428 : cluster [DBG] pgmap v343: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:39.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:38 vm07 bash[19945]: cluster 2026-03-21T06:57:37.622148+0000 mgr.x (mgr.14152) 428 : cluster [DBG] pgmap v343: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:40 vm04 bash[20194]: cluster 2026-03-21T06:57:39.622555+0000 mgr.x (mgr.14152) 429 : cluster [DBG] pgmap v344: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:41.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:40 vm04 bash[20194]: cluster 2026-03-21T06:57:39.622555+0000 mgr.x (mgr.14152) 429 : cluster [DBG] pgmap v344: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:41.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:40 vm02 bash[17657]: cluster 2026-03-21T06:57:39.622555+0000 mgr.x (mgr.14152) 429 : cluster [DBG] pgmap v344: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:41.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:40 vm02 bash[17657]: cluster 2026-03-21T06:57:39.622555+0000 mgr.x (mgr.14152) 429 : cluster [DBG] pgmap v344: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:40 vm07 bash[19945]: cluster 2026-03-21T06:57:39.622555+0000 mgr.x (mgr.14152) 429 : cluster [DBG] pgmap v344: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:41.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:40 vm07 bash[19945]: cluster 2026-03-21T06:57:39.622555+0000 mgr.x (mgr.14152) 429 : cluster [DBG] pgmap v344: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:43.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:42 vm04 bash[20194]: cluster 2026-03-21T06:57:41.623037+0000 mgr.x (mgr.14152) 430 : cluster [DBG] pgmap v345: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:43.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:42 vm04 bash[20194]: cluster 2026-03-21T06:57:41.623037+0000 mgr.x (mgr.14152) 430 : cluster [DBG] pgmap v345: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:43.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:42 vm02 bash[17657]: cluster 2026-03-21T06:57:41.623037+0000 mgr.x (mgr.14152) 430 : cluster [DBG] pgmap v345: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:43.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:42 vm02 bash[17657]: cluster 2026-03-21T06:57:41.623037+0000 mgr.x (mgr.14152) 430 : cluster [DBG] pgmap v345: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:42 vm07 bash[19945]: cluster 2026-03-21T06:57:41.623037+0000 mgr.x (mgr.14152) 430 : cluster [DBG] pgmap v345: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:43.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:42 vm07 bash[19945]: cluster 2026-03-21T06:57:41.623037+0000 mgr.x (mgr.14152) 430 : cluster [DBG] pgmap v345: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:44 vm04 bash[20194]: cluster 2026-03-21T06:57:43.623411+0000 mgr.x (mgr.14152) 431 : cluster [DBG] pgmap v346: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:45.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:44 vm04 bash[20194]: cluster 2026-03-21T06:57:43.623411+0000 mgr.x (mgr.14152) 431 : cluster [DBG] pgmap v346: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:45.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:44 vm02 bash[17657]: cluster 2026-03-21T06:57:43.623411+0000 mgr.x (mgr.14152) 431 : cluster [DBG] pgmap v346: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:45.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:44 vm02 bash[17657]: cluster 2026-03-21T06:57:43.623411+0000 mgr.x (mgr.14152) 431 : cluster [DBG] pgmap v346: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:45.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:44 vm07 bash[19945]: cluster 2026-03-21T06:57:43.623411+0000 mgr.x (mgr.14152) 431 : cluster [DBG] pgmap v346: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:45.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:44 vm07 bash[19945]: cluster 2026-03-21T06:57:43.623411+0000 mgr.x (mgr.14152) 431 : cluster [DBG] pgmap v346: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:46 vm04 bash[20194]: cluster 2026-03-21T06:57:45.623781+0000 mgr.x (mgr.14152) 432 : cluster [DBG] pgmap v347: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:47.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:46 vm04 bash[20194]: cluster 2026-03-21T06:57:45.623781+0000 mgr.x (mgr.14152) 432 : cluster [DBG] pgmap v347: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:47.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:57:46 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:57:47.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:46 vm02 bash[17657]: cluster 2026-03-21T06:57:45.623781+0000 mgr.x (mgr.14152) 432 : cluster [DBG] pgmap v347: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:47.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:46 vm02 bash[17657]: cluster 2026-03-21T06:57:45.623781+0000 mgr.x (mgr.14152) 432 : cluster [DBG] pgmap v347: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:47.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:46 vm07 bash[19945]: cluster 2026-03-21T06:57:45.623781+0000 mgr.x (mgr.14152) 432 : cluster [DBG] pgmap v347: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:47.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:46 vm07 bash[19945]: cluster 2026-03-21T06:57:45.623781+0000 mgr.x (mgr.14152) 432 : cluster [DBG] pgmap v347: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:47 vm07 bash[19945]: audit 2026-03-21T06:57:46.711367+0000 mgr.x (mgr.14152) 433 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:47.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:47 vm07 bash[19945]: audit 2026-03-21T06:57:46.711367+0000 mgr.x (mgr.14152) 433 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:47.901 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:47 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:57:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:47 vm04 bash[20194]: audit 2026-03-21T06:57:46.711367+0000 mgr.x (mgr.14152) 433 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:48.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:47 vm04 bash[20194]: audit 2026-03-21T06:57:46.711367+0000 mgr.x (mgr.14152) 433 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:48.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:47 vm02 bash[17657]: audit 2026-03-21T06:57:46.711367+0000 mgr.x (mgr.14152) 433 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:48.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:47 vm02 bash[17657]: audit 2026-03-21T06:57:46.711367+0000 mgr.x (mgr.14152) 433 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:48 vm04 bash[20194]: audit 2026-03-21T06:57:47.546770+0000 mgr.x (mgr.14152) 434 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:48 vm04 bash[20194]: audit 2026-03-21T06:57:47.546770+0000 mgr.x (mgr.14152) 434 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:48 vm04 bash[20194]: cluster 2026-03-21T06:57:47.624134+0000 mgr.x (mgr.14152) 435 : cluster [DBG] pgmap v348: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:49.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:48 vm04 bash[20194]: cluster 2026-03-21T06:57:47.624134+0000 mgr.x (mgr.14152) 435 : cluster [DBG] pgmap v348: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:48 vm02 bash[17657]: audit 2026-03-21T06:57:47.546770+0000 mgr.x (mgr.14152) 434 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:48 vm02 bash[17657]: audit 2026-03-21T06:57:47.546770+0000 mgr.x (mgr.14152) 434 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:48 vm02 bash[17657]: cluster 2026-03-21T06:57:47.624134+0000 mgr.x (mgr.14152) 435 : cluster [DBG] pgmap v348: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:48 vm02 bash[17657]: cluster 2026-03-21T06:57:47.624134+0000 mgr.x (mgr.14152) 435 : cluster [DBG] pgmap v348: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:49.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:48 vm07 bash[19945]: audit 2026-03-21T06:57:47.546770+0000 mgr.x (mgr.14152) 434 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:49.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:48 vm07 bash[19945]: audit 2026-03-21T06:57:47.546770+0000 mgr.x (mgr.14152) 434 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:49.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:48 vm07 bash[19945]: cluster 2026-03-21T06:57:47.624134+0000 mgr.x (mgr.14152) 435 : cluster [DBG] pgmap v348: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:49.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:48 vm07 bash[19945]: cluster 2026-03-21T06:57:47.624134+0000 mgr.x (mgr.14152) 435 : cluster [DBG] pgmap v348: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:50 vm04 bash[20194]: cluster 2026-03-21T06:57:49.624570+0000 mgr.x (mgr.14152) 436 : cluster [DBG] pgmap v349: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:51.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:50 vm04 bash[20194]: cluster 2026-03-21T06:57:49.624570+0000 mgr.x (mgr.14152) 436 : cluster [DBG] pgmap v349: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:51.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:50 vm02 bash[17657]: cluster 2026-03-21T06:57:49.624570+0000 mgr.x (mgr.14152) 436 : cluster [DBG] pgmap v349: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:51.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:50 vm02 bash[17657]: cluster 2026-03-21T06:57:49.624570+0000 mgr.x (mgr.14152) 436 : cluster [DBG] pgmap v349: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:50 vm07 bash[19945]: cluster 2026-03-21T06:57:49.624570+0000 mgr.x (mgr.14152) 436 : cluster [DBG] pgmap v349: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:50 vm07 bash[19945]: cluster 2026-03-21T06:57:49.624570+0000 mgr.x (mgr.14152) 436 : cluster [DBG] pgmap v349: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:53.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:52 vm04 bash[20194]: cluster 2026-03-21T06:57:51.624996+0000 mgr.x (mgr.14152) 437 : cluster [DBG] pgmap v350: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:53.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:52 vm04 bash[20194]: cluster 2026-03-21T06:57:51.624996+0000 mgr.x (mgr.14152) 437 : cluster [DBG] pgmap v350: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:53.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:52 vm02 bash[17657]: cluster 2026-03-21T06:57:51.624996+0000 mgr.x (mgr.14152) 437 : cluster [DBG] pgmap v350: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:53.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:52 vm02 bash[17657]: cluster 2026-03-21T06:57:51.624996+0000 mgr.x (mgr.14152) 437 : cluster [DBG] pgmap v350: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:52 vm07 bash[19945]: cluster 2026-03-21T06:57:51.624996+0000 mgr.x (mgr.14152) 437 : cluster [DBG] pgmap v350: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:53.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:52 vm07 bash[19945]: cluster 2026-03-21T06:57:51.624996+0000 mgr.x (mgr.14152) 437 : cluster [DBG] pgmap v350: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr:Note: switching to '0392f78529848ec72469e8e431875cb98d3a5fb4'. 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr:state without impacting any branches by switching back to a branch. 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr: git switch -c 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr:Or undo this operation with: 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr: git switch - 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T06:57:53.496 INFO:teuthology.orchestra.run.vm02.stderr:HEAD is now at 0392f785298 qa/tasks/keystone: restart mariadb for rocky and alma linux too 2026-03-21T06:57:53.502 DEBUG:teuthology.orchestra.run.vm02:> cp -- /home/ubuntu/cephtest/clone.client.0/src/test/cli-integration/rbd/gwcli_create.t /home/ubuntu/cephtest/archive/cram.client.0 2026-03-21T06:57:53.546 DEBUG:teuthology.orchestra.run.vm04:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.1 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-21T06:57:54.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:53 vm02 bash[17657]: cluster 2026-03-21T06:57:53.625372+0000 mgr.x (mgr.14152) 438 : cluster [DBG] pgmap v351: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:54.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:53 vm02 bash[17657]: cluster 2026-03-21T06:57:53.625372+0000 mgr.x (mgr.14152) 438 : cluster [DBG] pgmap v351: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:53 vm07 bash[19945]: cluster 2026-03-21T06:57:53.625372+0000 mgr.x (mgr.14152) 438 : cluster [DBG] pgmap v351: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:54.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:53 vm07 bash[19945]: cluster 2026-03-21T06:57:53.625372+0000 mgr.x (mgr.14152) 438 : cluster [DBG] pgmap v351: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:54.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:53 vm04 bash[20194]: cluster 2026-03-21T06:57:53.625372+0000 mgr.x (mgr.14152) 438 : cluster [DBG] pgmap v351: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:54.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:53 vm04 bash[20194]: cluster 2026-03-21T06:57:53.625372+0000 mgr.x (mgr.14152) 438 : cluster [DBG] pgmap v351: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:55.228 INFO:teuthology.orchestra.run.vm04.stdout:Collecting cram==0.6 2026-03-21T06:57:55.268 INFO:teuthology.orchestra.run.vm04.stdout: Downloading cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-21T06:57:55.283 INFO:teuthology.orchestra.run.vm04.stdout:Installing collected packages: cram 2026-03-21T06:57:55.288 INFO:teuthology.orchestra.run.vm04.stdout:Successfully installed cram-0.6 2026-03-21T06:57:55.318 DEBUG:teuthology.orchestra.run.vm04:> rm -rf /home/ubuntu/cephtest/clone.client.1 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.1 && cd /home/ubuntu/cephtest/clone.client.1 && git checkout 0392f78529848ec72469e8e431875cb98d3a5fb4 2026-03-21T06:57:55.322 INFO:teuthology.orchestra.run.vm04.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.1'... 2026-03-21T06:57:57.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:56 vm04 bash[20194]: cluster 2026-03-21T06:57:55.625733+0000 mgr.x (mgr.14152) 439 : cluster [DBG] pgmap v352: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:57.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:56 vm04 bash[20194]: cluster 2026-03-21T06:57:55.625733+0000 mgr.x (mgr.14152) 439 : cluster [DBG] pgmap v352: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:57.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:57:56 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:57:57.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:56 vm02 bash[17657]: cluster 2026-03-21T06:57:55.625733+0000 mgr.x (mgr.14152) 439 : cluster [DBG] pgmap v352: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:57.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:56 vm02 bash[17657]: cluster 2026-03-21T06:57:55.625733+0000 mgr.x (mgr.14152) 439 : cluster [DBG] pgmap v352: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:57.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:56 vm07 bash[19945]: cluster 2026-03-21T06:57:55.625733+0000 mgr.x (mgr.14152) 439 : cluster [DBG] pgmap v352: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:57.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:56 vm07 bash[19945]: cluster 2026-03-21T06:57:55.625733+0000 mgr.x (mgr.14152) 439 : cluster [DBG] pgmap v352: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:57:57.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:57 vm07 bash[19945]: audit 2026-03-21T06:57:56.719424+0000 mgr.x (mgr.14152) 440 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:57.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:57 vm07 bash[19945]: audit 2026-03-21T06:57:56.719424+0000 mgr.x (mgr.14152) 440 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:57.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:57:57 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:57:58.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:57 vm04 bash[20194]: audit 2026-03-21T06:57:56.719424+0000 mgr.x (mgr.14152) 440 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:58.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:57 vm04 bash[20194]: audit 2026-03-21T06:57:56.719424+0000 mgr.x (mgr.14152) 440 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:58.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:57 vm02 bash[17657]: audit 2026-03-21T06:57:56.719424+0000 mgr.x (mgr.14152) 440 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:58.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:57 vm02 bash[17657]: audit 2026-03-21T06:57:56.719424+0000 mgr.x (mgr.14152) 440 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:59.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:58 vm04 bash[20194]: audit 2026-03-21T06:57:57.557719+0000 mgr.x (mgr.14152) 441 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:59.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:58 vm04 bash[20194]: audit 2026-03-21T06:57:57.557719+0000 mgr.x (mgr.14152) 441 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:59.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:58 vm04 bash[20194]: cluster 2026-03-21T06:57:57.626068+0000 mgr.x (mgr.14152) 442 : cluster [DBG] pgmap v353: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:59.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:57:58 vm04 bash[20194]: cluster 2026-03-21T06:57:57.626068+0000 mgr.x (mgr.14152) 442 : cluster [DBG] pgmap v353: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:58 vm02 bash[17657]: audit 2026-03-21T06:57:57.557719+0000 mgr.x (mgr.14152) 441 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:58 vm02 bash[17657]: audit 2026-03-21T06:57:57.557719+0000 mgr.x (mgr.14152) 441 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:58 vm02 bash[17657]: cluster 2026-03-21T06:57:57.626068+0000 mgr.x (mgr.14152) 442 : cluster [DBG] pgmap v353: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:57:58 vm02 bash[17657]: cluster 2026-03-21T06:57:57.626068+0000 mgr.x (mgr.14152) 442 : cluster [DBG] pgmap v353: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:58 vm07 bash[19945]: audit 2026-03-21T06:57:57.557719+0000 mgr.x (mgr.14152) 441 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:58 vm07 bash[19945]: audit 2026-03-21T06:57:57.557719+0000 mgr.x (mgr.14152) 441 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:57:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:58 vm07 bash[19945]: cluster 2026-03-21T06:57:57.626068+0000 mgr.x (mgr.14152) 442 : cluster [DBG] pgmap v353: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:57:59.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:57:58 vm07 bash[19945]: cluster 2026-03-21T06:57:57.626068+0000 mgr.x (mgr.14152) 442 : cluster [DBG] pgmap v353: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:00 vm04 bash[20194]: cluster 2026-03-21T06:57:59.626469+0000 mgr.x (mgr.14152) 443 : cluster [DBG] pgmap v354: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:01.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:00 vm04 bash[20194]: cluster 2026-03-21T06:57:59.626469+0000 mgr.x (mgr.14152) 443 : cluster [DBG] pgmap v354: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:01.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:00 vm02 bash[17657]: cluster 2026-03-21T06:57:59.626469+0000 mgr.x (mgr.14152) 443 : cluster [DBG] pgmap v354: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:01.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:00 vm02 bash[17657]: cluster 2026-03-21T06:57:59.626469+0000 mgr.x (mgr.14152) 443 : cluster [DBG] pgmap v354: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:01.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:00 vm07 bash[19945]: cluster 2026-03-21T06:57:59.626469+0000 mgr.x (mgr.14152) 443 : cluster [DBG] pgmap v354: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:01.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:00 vm07 bash[19945]: cluster 2026-03-21T06:57:59.626469+0000 mgr.x (mgr.14152) 443 : cluster [DBG] pgmap v354: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:02 vm04 bash[20194]: cluster 2026-03-21T06:58:01.626849+0000 mgr.x (mgr.14152) 444 : cluster [DBG] pgmap v355: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:03.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:02 vm04 bash[20194]: cluster 2026-03-21T06:58:01.626849+0000 mgr.x (mgr.14152) 444 : cluster [DBG] pgmap v355: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:03.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:02 vm02 bash[17657]: cluster 2026-03-21T06:58:01.626849+0000 mgr.x (mgr.14152) 444 : cluster [DBG] pgmap v355: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:03.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:02 vm02 bash[17657]: cluster 2026-03-21T06:58:01.626849+0000 mgr.x (mgr.14152) 444 : cluster [DBG] pgmap v355: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:03.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:02 vm07 bash[19945]: cluster 2026-03-21T06:58:01.626849+0000 mgr.x (mgr.14152) 444 : cluster [DBG] pgmap v355: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:03.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:02 vm07 bash[19945]: cluster 2026-03-21T06:58:01.626849+0000 mgr.x (mgr.14152) 444 : cluster [DBG] pgmap v355: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:04.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:03 vm02 bash[17657]: cluster 2026-03-21T06:58:03.627173+0000 mgr.x (mgr.14152) 445 : cluster [DBG] pgmap v356: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:04.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:03 vm02 bash[17657]: cluster 2026-03-21T06:58:03.627173+0000 mgr.x (mgr.14152) 445 : cluster [DBG] pgmap v356: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:04.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:03 vm07 bash[19945]: cluster 2026-03-21T06:58:03.627173+0000 mgr.x (mgr.14152) 445 : cluster [DBG] pgmap v356: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:04.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:03 vm07 bash[19945]: cluster 2026-03-21T06:58:03.627173+0000 mgr.x (mgr.14152) 445 : cluster [DBG] pgmap v356: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:04.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:03 vm04 bash[20194]: cluster 2026-03-21T06:58:03.627173+0000 mgr.x (mgr.14152) 445 : cluster [DBG] pgmap v356: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:04.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:03 vm04 bash[20194]: cluster 2026-03-21T06:58:03.627173+0000 mgr.x (mgr.14152) 445 : cluster [DBG] pgmap v356: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:07.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:06 vm04 bash[20194]: cluster 2026-03-21T06:58:05.627485+0000 mgr.x (mgr.14152) 446 : cluster [DBG] pgmap v357: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:07.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:06 vm04 bash[20194]: cluster 2026-03-21T06:58:05.627485+0000 mgr.x (mgr.14152) 446 : cluster [DBG] pgmap v357: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:07.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:06 vm02 bash[17657]: cluster 2026-03-21T06:58:05.627485+0000 mgr.x (mgr.14152) 446 : cluster [DBG] pgmap v357: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:07.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:06 vm02 bash[17657]: cluster 2026-03-21T06:58:05.627485+0000 mgr.x (mgr.14152) 446 : cluster [DBG] pgmap v357: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:07.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:58:06 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:58:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:06 vm07 bash[19945]: cluster 2026-03-21T06:58:05.627485+0000 mgr.x (mgr.14152) 446 : cluster [DBG] pgmap v357: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:07.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:06 vm07 bash[19945]: cluster 2026-03-21T06:58:05.627485+0000 mgr.x (mgr.14152) 446 : cluster [DBG] pgmap v357: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:07 vm07 bash[19945]: audit 2026-03-21T06:58:06.723387+0000 mgr.x (mgr.14152) 447 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:07.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:07 vm07 bash[19945]: audit 2026-03-21T06:58:06.723387+0000 mgr.x (mgr.14152) 447 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:07.901 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:58:07 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:58:08.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:07 vm04 bash[20194]: audit 2026-03-21T06:58:06.723387+0000 mgr.x (mgr.14152) 447 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:08.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:07 vm04 bash[20194]: audit 2026-03-21T06:58:06.723387+0000 mgr.x (mgr.14152) 447 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:08.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:07 vm02 bash[17657]: audit 2026-03-21T06:58:06.723387+0000 mgr.x (mgr.14152) 447 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:08.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:07 vm02 bash[17657]: audit 2026-03-21T06:58:06.723387+0000 mgr.x (mgr.14152) 447 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:09.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:09 vm02 bash[17657]: audit 2026-03-21T06:58:07.568161+0000 mgr.x (mgr.14152) 448 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:09.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:09 vm02 bash[17657]: audit 2026-03-21T06:58:07.568161+0000 mgr.x (mgr.14152) 448 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:09.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:09 vm02 bash[17657]: cluster 2026-03-21T06:58:07.627745+0000 mgr.x (mgr.14152) 449 : cluster [DBG] pgmap v358: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:09.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:09 vm02 bash[17657]: cluster 2026-03-21T06:58:07.627745+0000 mgr.x (mgr.14152) 449 : cluster [DBG] pgmap v358: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:09.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:09 vm07 bash[19945]: audit 2026-03-21T06:58:07.568161+0000 mgr.x (mgr.14152) 448 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:09.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:09 vm07 bash[19945]: audit 2026-03-21T06:58:07.568161+0000 mgr.x (mgr.14152) 448 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:09.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:09 vm07 bash[19945]: cluster 2026-03-21T06:58:07.627745+0000 mgr.x (mgr.14152) 449 : cluster [DBG] pgmap v358: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:09.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:09 vm07 bash[19945]: cluster 2026-03-21T06:58:07.627745+0000 mgr.x (mgr.14152) 449 : cluster [DBG] pgmap v358: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:09.508 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:09 vm04 bash[20194]: audit 2026-03-21T06:58:07.568161+0000 mgr.x (mgr.14152) 448 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:09.508 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:09 vm04 bash[20194]: audit 2026-03-21T06:58:07.568161+0000 mgr.x (mgr.14152) 448 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:09.508 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:09 vm04 bash[20194]: cluster 2026-03-21T06:58:07.627745+0000 mgr.x (mgr.14152) 449 : cluster [DBG] pgmap v358: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:09.508 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:09 vm04 bash[20194]: cluster 2026-03-21T06:58:07.627745+0000 mgr.x (mgr.14152) 449 : cluster [DBG] pgmap v358: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:10.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:10 vm02 bash[17657]: cluster 2026-03-21T06:58:09.628037+0000 mgr.x (mgr.14152) 450 : cluster [DBG] pgmap v359: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:10.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:10 vm02 bash[17657]: cluster 2026-03-21T06:58:09.628037+0000 mgr.x (mgr.14152) 450 : cluster [DBG] pgmap v359: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:10.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:10 vm07 bash[19945]: cluster 2026-03-21T06:58:09.628037+0000 mgr.x (mgr.14152) 450 : cluster [DBG] pgmap v359: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:10.401 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:10 vm07 bash[19945]: cluster 2026-03-21T06:58:09.628037+0000 mgr.x (mgr.14152) 450 : cluster [DBG] pgmap v359: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:10.508 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:10 vm04 bash[20194]: cluster 2026-03-21T06:58:09.628037+0000 mgr.x (mgr.14152) 450 : cluster [DBG] pgmap v359: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:10.508 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:10 vm04 bash[20194]: cluster 2026-03-21T06:58:09.628037+0000 mgr.x (mgr.14152) 450 : cluster [DBG] pgmap v359: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:13.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:12 vm04 bash[20194]: cluster 2026-03-21T06:58:11.628358+0000 mgr.x (mgr.14152) 451 : cluster [DBG] pgmap v360: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:13.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:12 vm04 bash[20194]: cluster 2026-03-21T06:58:11.628358+0000 mgr.x (mgr.14152) 451 : cluster [DBG] pgmap v360: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:13.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:12 vm02 bash[17657]: cluster 2026-03-21T06:58:11.628358+0000 mgr.x (mgr.14152) 451 : cluster [DBG] pgmap v360: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:13.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:12 vm02 bash[17657]: cluster 2026-03-21T06:58:11.628358+0000 mgr.x (mgr.14152) 451 : cluster [DBG] pgmap v360: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:12 vm07 bash[19945]: cluster 2026-03-21T06:58:11.628358+0000 mgr.x (mgr.14152) 451 : cluster [DBG] pgmap v360: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:13.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:12 vm07 bash[19945]: cluster 2026-03-21T06:58:11.628358+0000 mgr.x (mgr.14152) 451 : cluster [DBG] pgmap v360: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:15.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:14 vm04 bash[20194]: cluster 2026-03-21T06:58:13.628653+0000 mgr.x (mgr.14152) 452 : cluster [DBG] pgmap v361: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:15.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:14 vm04 bash[20194]: cluster 2026-03-21T06:58:13.628653+0000 mgr.x (mgr.14152) 452 : cluster [DBG] pgmap v361: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:15.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:14 vm02 bash[17657]: cluster 2026-03-21T06:58:13.628653+0000 mgr.x (mgr.14152) 452 : cluster [DBG] pgmap v361: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:15.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:14 vm02 bash[17657]: cluster 2026-03-21T06:58:13.628653+0000 mgr.x (mgr.14152) 452 : cluster [DBG] pgmap v361: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:14 vm07 bash[19945]: cluster 2026-03-21T06:58:13.628653+0000 mgr.x (mgr.14152) 452 : cluster [DBG] pgmap v361: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:15.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:14 vm07 bash[19945]: cluster 2026-03-21T06:58:13.628653+0000 mgr.x (mgr.14152) 452 : cluster [DBG] pgmap v361: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:17.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:16 vm04 bash[20194]: cluster 2026-03-21T06:58:15.628981+0000 mgr.x (mgr.14152) 453 : cluster [DBG] pgmap v362: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:16 vm04 bash[20194]: cluster 2026-03-21T06:58:15.628981+0000 mgr.x (mgr.14152) 453 : cluster [DBG] pgmap v362: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:16 vm04 bash[20194]: audit 2026-03-21T06:58:16.295138+0000 mon.a (mon.0) 863 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:58:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:16 vm04 bash[20194]: audit 2026-03-21T06:58:16.295138+0000 mon.a (mon.0) 863 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:58:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:16 vm04 bash[20194]: audit 2026-03-21T06:58:16.612949+0000 mon.a (mon.0) 864 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:16 vm04 bash[20194]: audit 2026-03-21T06:58:16.612949+0000 mon.a (mon.0) 864 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:16 vm04 bash[20194]: audit 2026-03-21T06:58:16.618149+0000 mon.a (mon.0) 865 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:16 vm04 bash[20194]: audit 2026-03-21T06:58:16.618149+0000 mon.a (mon.0) 865 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:58:16 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:58:17.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:16 vm02 bash[17657]: cluster 2026-03-21T06:58:15.628981+0000 mgr.x (mgr.14152) 453 : cluster [DBG] pgmap v362: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:17.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:16 vm02 bash[17657]: cluster 2026-03-21T06:58:15.628981+0000 mgr.x (mgr.14152) 453 : cluster [DBG] pgmap v362: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:17.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:16 vm02 bash[17657]: audit 2026-03-21T06:58:16.295138+0000 mon.a (mon.0) 863 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:58:17.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:16 vm02 bash[17657]: audit 2026-03-21T06:58:16.295138+0000 mon.a (mon.0) 863 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:58:17.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:16 vm02 bash[17657]: audit 2026-03-21T06:58:16.612949+0000 mon.a (mon.0) 864 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:16 vm02 bash[17657]: audit 2026-03-21T06:58:16.612949+0000 mon.a (mon.0) 864 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:16 vm02 bash[17657]: audit 2026-03-21T06:58:16.618149+0000 mon.a (mon.0) 865 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:16 vm02 bash[17657]: audit 2026-03-21T06:58:16.618149+0000 mon.a (mon.0) 865 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:16 vm07 bash[19945]: cluster 2026-03-21T06:58:15.628981+0000 mgr.x (mgr.14152) 453 : cluster [DBG] pgmap v362: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:16 vm07 bash[19945]: cluster 2026-03-21T06:58:15.628981+0000 mgr.x (mgr.14152) 453 : cluster [DBG] pgmap v362: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:16 vm07 bash[19945]: audit 2026-03-21T06:58:16.295138+0000 mon.a (mon.0) 863 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:58:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:16 vm07 bash[19945]: audit 2026-03-21T06:58:16.295138+0000 mon.a (mon.0) 863 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:58:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:16 vm07 bash[19945]: audit 2026-03-21T06:58:16.612949+0000 mon.a (mon.0) 864 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:16 vm07 bash[19945]: audit 2026-03-21T06:58:16.612949+0000 mon.a (mon.0) 864 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:16 vm07 bash[19945]: audit 2026-03-21T06:58:16.618149+0000 mon.a (mon.0) 865 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:16 vm07 bash[19945]: audit 2026-03-21T06:58:16.618149+0000 mon.a (mon.0) 865 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:17.901 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:17 vm07 bash[19945]: audit 2026-03-21T06:58:16.731379+0000 mgr.x (mgr.14152) 454 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:17.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:17 vm07 bash[19945]: audit 2026-03-21T06:58:16.731379+0000 mgr.x (mgr.14152) 454 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:17.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:58:17 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:58:18.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:17 vm04 bash[20194]: audit 2026-03-21T06:58:16.731379+0000 mgr.x (mgr.14152) 454 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:18.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:17 vm04 bash[20194]: audit 2026-03-21T06:58:16.731379+0000 mgr.x (mgr.14152) 454 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:18.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:17 vm02 bash[17657]: audit 2026-03-21T06:58:16.731379+0000 mgr.x (mgr.14152) 454 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:18.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:17 vm02 bash[17657]: audit 2026-03-21T06:58:16.731379+0000 mgr.x (mgr.14152) 454 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:19.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:18 vm04 bash[20194]: audit 2026-03-21T06:58:17.578884+0000 mgr.x (mgr.14152) 455 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:18 vm04 bash[20194]: audit 2026-03-21T06:58:17.578884+0000 mgr.x (mgr.14152) 455 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:18 vm04 bash[20194]: cluster 2026-03-21T06:58:17.629248+0000 mgr.x (mgr.14152) 456 : cluster [DBG] pgmap v363: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:18 vm04 bash[20194]: cluster 2026-03-21T06:58:17.629248+0000 mgr.x (mgr.14152) 456 : cluster [DBG] pgmap v363: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:19.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:18 vm02 bash[17657]: audit 2026-03-21T06:58:17.578884+0000 mgr.x (mgr.14152) 455 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:19.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:18 vm02 bash[17657]: audit 2026-03-21T06:58:17.578884+0000 mgr.x (mgr.14152) 455 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:19.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:18 vm02 bash[17657]: cluster 2026-03-21T06:58:17.629248+0000 mgr.x (mgr.14152) 456 : cluster [DBG] pgmap v363: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:19.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:18 vm02 bash[17657]: cluster 2026-03-21T06:58:17.629248+0000 mgr.x (mgr.14152) 456 : cluster [DBG] pgmap v363: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:18 vm07 bash[19945]: audit 2026-03-21T06:58:17.578884+0000 mgr.x (mgr.14152) 455 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:18 vm07 bash[19945]: audit 2026-03-21T06:58:17.578884+0000 mgr.x (mgr.14152) 455 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:18 vm07 bash[19945]: cluster 2026-03-21T06:58:17.629248+0000 mgr.x (mgr.14152) 456 : cluster [DBG] pgmap v363: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:19.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:18 vm07 bash[19945]: cluster 2026-03-21T06:58:17.629248+0000 mgr.x (mgr.14152) 456 : cluster [DBG] pgmap v363: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:21.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:20 vm04 bash[20194]: cluster 2026-03-21T06:58:19.629550+0000 mgr.x (mgr.14152) 457 : cluster [DBG] pgmap v364: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:20 vm04 bash[20194]: cluster 2026-03-21T06:58:19.629550+0000 mgr.x (mgr.14152) 457 : cluster [DBG] pgmap v364: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:20 vm02 bash[17657]: cluster 2026-03-21T06:58:19.629550+0000 mgr.x (mgr.14152) 457 : cluster [DBG] pgmap v364: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:20 vm02 bash[17657]: cluster 2026-03-21T06:58:19.629550+0000 mgr.x (mgr.14152) 457 : cluster [DBG] pgmap v364: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:20 vm07 bash[19945]: cluster 2026-03-21T06:58:19.629550+0000 mgr.x (mgr.14152) 457 : cluster [DBG] pgmap v364: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:21.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:20 vm07 bash[19945]: cluster 2026-03-21T06:58:19.629550+0000 mgr.x (mgr.14152) 457 : cluster [DBG] pgmap v364: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.354697+0000 mon.a (mon.0) 866 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.354697+0000 mon.a (mon.0) 866 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.359539+0000 mon.a (mon.0) 867 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.359539+0000 mon.a (mon.0) 867 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: cluster 2026-03-21T06:58:21.629906+0000 mgr.x (mgr.14152) 458 : cluster [DBG] pgmap v365: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: cluster 2026-03-21T06:58:21.629906+0000 mgr.x (mgr.14152) 458 : cluster [DBG] pgmap v365: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.669098+0000 mon.a (mon.0) 868 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.669098+0000 mon.a (mon.0) 868 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.669710+0000 mon.a (mon.0) 869 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.669710+0000 mon.a (mon.0) 869 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.674892+0000 mon.a (mon.0) 870 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.674892+0000 mon.a (mon.0) 870 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.675811+0000 mon.a (mon.0) 871 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:58:22.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:22 vm02 bash[17657]: audit 2026-03-21T06:58:21.675811+0000 mon.a (mon.0) 871 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:58:22.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.354697+0000 mon.a (mon.0) 866 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.354697+0000 mon.a (mon.0) 866 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.359539+0000 mon.a (mon.0) 867 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.359539+0000 mon.a (mon.0) 867 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.651 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: cluster 2026-03-21T06:58:21.629906+0000 mgr.x (mgr.14152) 458 : cluster [DBG] pgmap v365: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:22.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: cluster 2026-03-21T06:58:21.629906+0000 mgr.x (mgr.14152) 458 : cluster [DBG] pgmap v365: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:22.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.669098+0000 mon.a (mon.0) 868 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:58:22.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.669098+0000 mon.a (mon.0) 868 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:58:22.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.669710+0000 mon.a (mon.0) 869 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:58:22.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.669710+0000 mon.a (mon.0) 869 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:58:22.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.674892+0000 mon.a (mon.0) 870 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.674892+0000 mon.a (mon.0) 870 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.675811+0000 mon.a (mon.0) 871 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:58:22.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:22 vm07 bash[19945]: audit 2026-03-21T06:58:21.675811+0000 mon.a (mon.0) 871 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:58:22.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.354697+0000 mon.a (mon.0) 866 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.354697+0000 mon.a (mon.0) 866 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.359539+0000 mon.a (mon.0) 867 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.359539+0000 mon.a (mon.0) 867 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: cluster 2026-03-21T06:58:21.629906+0000 mgr.x (mgr.14152) 458 : cluster [DBG] pgmap v365: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:22.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: cluster 2026-03-21T06:58:21.629906+0000 mgr.x (mgr.14152) 458 : cluster [DBG] pgmap v365: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:22.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.669098+0000 mon.a (mon.0) 868 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:58:22.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.669098+0000 mon.a (mon.0) 868 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:58:22.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.669710+0000 mon.a (mon.0) 869 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:58:22.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.669710+0000 mon.a (mon.0) 869 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:58:22.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.674892+0000 mon.a (mon.0) 870 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.674892+0000 mon.a (mon.0) 870 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:58:22.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.675811+0000 mon.a (mon.0) 871 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:58:22.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:22 vm04 bash[20194]: audit 2026-03-21T06:58:21.675811+0000 mon.a (mon.0) 871 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:58:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:23 vm02 bash[17657]: cluster 2026-03-21T06:58:23.630258+0000 mgr.x (mgr.14152) 459 : cluster [DBG] pgmap v366: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:23 vm02 bash[17657]: cluster 2026-03-21T06:58:23.630258+0000 mgr.x (mgr.14152) 459 : cluster [DBG] pgmap v366: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:24.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:23 vm07 bash[19945]: cluster 2026-03-21T06:58:23.630258+0000 mgr.x (mgr.14152) 459 : cluster [DBG] pgmap v366: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:23 vm07 bash[19945]: cluster 2026-03-21T06:58:23.630258+0000 mgr.x (mgr.14152) 459 : cluster [DBG] pgmap v366: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:24.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:23 vm04 bash[20194]: cluster 2026-03-21T06:58:23.630258+0000 mgr.x (mgr.14152) 459 : cluster [DBG] pgmap v366: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:24.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:23 vm04 bash[20194]: cluster 2026-03-21T06:58:23.630258+0000 mgr.x (mgr.14152) 459 : cluster [DBG] pgmap v366: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:27.008 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:26 vm04 bash[20194]: cluster 2026-03-21T06:58:25.630614+0000 mgr.x (mgr.14152) 460 : cluster [DBG] pgmap v367: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:27.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:26 vm04 bash[20194]: cluster 2026-03-21T06:58:25.630614+0000 mgr.x (mgr.14152) 460 : cluster [DBG] pgmap v367: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:58:26 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:58:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:26 vm02 bash[17657]: cluster 2026-03-21T06:58:25.630614+0000 mgr.x (mgr.14152) 460 : cluster [DBG] pgmap v367: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:26 vm02 bash[17657]: cluster 2026-03-21T06:58:25.630614+0000 mgr.x (mgr.14152) 460 : cluster [DBG] pgmap v367: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:26 vm07 bash[19945]: cluster 2026-03-21T06:58:25.630614+0000 mgr.x (mgr.14152) 460 : cluster [DBG] pgmap v367: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:27.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:26 vm07 bash[19945]: cluster 2026-03-21T06:58:25.630614+0000 mgr.x (mgr.14152) 460 : cluster [DBG] pgmap v367: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:27.840 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:58:27 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:58:28.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:27 vm02 bash[17657]: audit 2026-03-21T06:58:26.741951+0000 mgr.x (mgr.14152) 461 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:28.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:27 vm02 bash[17657]: audit 2026-03-21T06:58:26.741951+0000 mgr.x (mgr.14152) 461 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:28.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:27 vm07 bash[19945]: audit 2026-03-21T06:58:26.741951+0000 mgr.x (mgr.14152) 461 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:28.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:27 vm07 bash[19945]: audit 2026-03-21T06:58:26.741951+0000 mgr.x (mgr.14152) 461 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:28.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:27 vm04 bash[20194]: audit 2026-03-21T06:58:26.741951+0000 mgr.x (mgr.14152) 461 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:28.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:27 vm04 bash[20194]: audit 2026-03-21T06:58:26.741951+0000 mgr.x (mgr.14152) 461 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:28 vm02 bash[17657]: audit 2026-03-21T06:58:27.589742+0000 mgr.x (mgr.14152) 462 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:28 vm02 bash[17657]: audit 2026-03-21T06:58:27.589742+0000 mgr.x (mgr.14152) 462 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:28 vm02 bash[17657]: cluster 2026-03-21T06:58:27.630891+0000 mgr.x (mgr.14152) 463 : cluster [DBG] pgmap v368: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:28 vm02 bash[17657]: cluster 2026-03-21T06:58:27.630891+0000 mgr.x (mgr.14152) 463 : cluster [DBG] pgmap v368: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:28 vm07 bash[19945]: audit 2026-03-21T06:58:27.589742+0000 mgr.x (mgr.14152) 462 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:29.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:28 vm07 bash[19945]: audit 2026-03-21T06:58:27.589742+0000 mgr.x (mgr.14152) 462 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:28 vm07 bash[19945]: cluster 2026-03-21T06:58:27.630891+0000 mgr.x (mgr.14152) 463 : cluster [DBG] pgmap v368: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:28 vm07 bash[19945]: cluster 2026-03-21T06:58:27.630891+0000 mgr.x (mgr.14152) 463 : cluster [DBG] pgmap v368: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:29.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:28 vm04 bash[20194]: audit 2026-03-21T06:58:27.589742+0000 mgr.x (mgr.14152) 462 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:29.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:28 vm04 bash[20194]: audit 2026-03-21T06:58:27.589742+0000 mgr.x (mgr.14152) 462 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:29.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:28 vm04 bash[20194]: cluster 2026-03-21T06:58:27.630891+0000 mgr.x (mgr.14152) 463 : cluster [DBG] pgmap v368: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:29.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:28 vm04 bash[20194]: cluster 2026-03-21T06:58:27.630891+0000 mgr.x (mgr.14152) 463 : cluster [DBG] pgmap v368: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:30.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:29 vm02 bash[17657]: cluster 2026-03-21T06:58:29.631199+0000 mgr.x (mgr.14152) 464 : cluster [DBG] pgmap v369: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:30.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:29 vm02 bash[17657]: cluster 2026-03-21T06:58:29.631199+0000 mgr.x (mgr.14152) 464 : cluster [DBG] pgmap v369: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:30.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:29 vm07 bash[19945]: cluster 2026-03-21T06:58:29.631199+0000 mgr.x (mgr.14152) 464 : cluster [DBG] pgmap v369: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:30.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:29 vm07 bash[19945]: cluster 2026-03-21T06:58:29.631199+0000 mgr.x (mgr.14152) 464 : cluster [DBG] pgmap v369: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:30.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:29 vm04 bash[20194]: cluster 2026-03-21T06:58:29.631199+0000 mgr.x (mgr.14152) 464 : cluster [DBG] pgmap v369: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:30.258 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:29 vm04 bash[20194]: cluster 2026-03-21T06:58:29.631199+0000 mgr.x (mgr.14152) 464 : cluster [DBG] pgmap v369: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:33.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:32 vm04 bash[20194]: cluster 2026-03-21T06:58:31.631512+0000 mgr.x (mgr.14152) 465 : cluster [DBG] pgmap v370: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:33.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:32 vm04 bash[20194]: cluster 2026-03-21T06:58:31.631512+0000 mgr.x (mgr.14152) 465 : cluster [DBG] pgmap v370: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:33.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:32 vm02 bash[17657]: cluster 2026-03-21T06:58:31.631512+0000 mgr.x (mgr.14152) 465 : cluster [DBG] pgmap v370: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:33.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:32 vm02 bash[17657]: cluster 2026-03-21T06:58:31.631512+0000 mgr.x (mgr.14152) 465 : cluster [DBG] pgmap v370: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:32 vm07 bash[19945]: cluster 2026-03-21T06:58:31.631512+0000 mgr.x (mgr.14152) 465 : cluster [DBG] pgmap v370: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:33.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:32 vm07 bash[19945]: cluster 2026-03-21T06:58:31.631512+0000 mgr.x (mgr.14152) 465 : cluster [DBG] pgmap v370: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:35.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:34 vm02 bash[17657]: cluster 2026-03-21T06:58:33.631828+0000 mgr.x (mgr.14152) 466 : cluster [DBG] pgmap v371: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:35.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:34 vm02 bash[17657]: cluster 2026-03-21T06:58:33.631828+0000 mgr.x (mgr.14152) 466 : cluster [DBG] pgmap v371: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:35.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:34 vm07 bash[19945]: cluster 2026-03-21T06:58:33.631828+0000 mgr.x (mgr.14152) 466 : cluster [DBG] pgmap v371: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:35.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:34 vm07 bash[19945]: cluster 2026-03-21T06:58:33.631828+0000 mgr.x (mgr.14152) 466 : cluster [DBG] pgmap v371: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr:Note: switching to '0392f78529848ec72469e8e431875cb98d3a5fb4'. 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr:state without impacting any branches by switching back to a branch. 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr: git switch -c 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr:Or undo this operation with: 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr: git switch - 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T06:58:35.178 INFO:teuthology.orchestra.run.vm04.stderr:HEAD is now at 0392f785298 qa/tasks/keystone: restart mariadb for rocky and alma linux too 2026-03-21T06:58:35.185 DEBUG:teuthology.orchestra.run.vm04:> cp -- /home/ubuntu/cephtest/clone.client.1/src/test/cli-integration/rbd/iscsi_client.t /home/ubuntu/cephtest/archive/cram.client.1 2026-03-21T06:58:35.229 DEBUG:teuthology.orchestra.run.vm07:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.2 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-21T06:58:35.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:34 vm04 bash[20194]: cluster 2026-03-21T06:58:33.631828+0000 mgr.x (mgr.14152) 466 : cluster [DBG] pgmap v371: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:35.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:34 vm04 bash[20194]: cluster 2026-03-21T06:58:33.631828+0000 mgr.x (mgr.14152) 466 : cluster [DBG] pgmap v371: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:36.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:35 vm02 bash[17657]: cluster 2026-03-21T06:58:35.632161+0000 mgr.x (mgr.14152) 467 : cluster [DBG] pgmap v372: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:36.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:35 vm02 bash[17657]: cluster 2026-03-21T06:58:35.632161+0000 mgr.x (mgr.14152) 467 : cluster [DBG] pgmap v372: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:36.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:35 vm07 bash[19945]: cluster 2026-03-21T06:58:35.632161+0000 mgr.x (mgr.14152) 467 : cluster [DBG] pgmap v372: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:36.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:35 vm07 bash[19945]: cluster 2026-03-21T06:58:35.632161+0000 mgr.x (mgr.14152) 467 : cluster [DBG] pgmap v372: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:36.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:35 vm04 bash[20194]: cluster 2026-03-21T06:58:35.632161+0000 mgr.x (mgr.14152) 467 : cluster [DBG] pgmap v372: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:36.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:35 vm04 bash[20194]: cluster 2026-03-21T06:58:35.632161+0000 mgr.x (mgr.14152) 467 : cluster [DBG] pgmap v372: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:37.079 INFO:teuthology.orchestra.run.vm07.stdout:Collecting cram==0.6 2026-03-21T06:58:37.122 INFO:teuthology.orchestra.run.vm07.stdout: Downloading cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-21T06:58:37.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:58:36 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:58:37.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:36 vm02 bash[17657]: audit 2026-03-21T06:58:36.752504+0000 mgr.x (mgr.14152) 468 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:37.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:36 vm02 bash[17657]: audit 2026-03-21T06:58:36.752504+0000 mgr.x (mgr.14152) 468 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:37.142 INFO:teuthology.orchestra.run.vm07.stdout:Installing collected packages: cram 2026-03-21T06:58:37.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:36 vm07 bash[19945]: audit 2026-03-21T06:58:36.752504+0000 mgr.x (mgr.14152) 468 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:37.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:36 vm07 bash[19945]: audit 2026-03-21T06:58:36.752504+0000 mgr.x (mgr.14152) 468 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:37.160 INFO:teuthology.orchestra.run.vm07.stdout:Successfully installed cram-0.6 2026-03-21T06:58:37.201 DEBUG:teuthology.orchestra.run.vm07:> rm -rf /home/ubuntu/cephtest/clone.client.2 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.2 && cd /home/ubuntu/cephtest/clone.client.2 && git checkout 0392f78529848ec72469e8e431875cb98d3a5fb4 2026-03-21T06:58:37.207 INFO:teuthology.orchestra.run.vm07.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.2'... 2026-03-21T06:58:37.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:36 vm04 bash[20194]: audit 2026-03-21T06:58:36.752504+0000 mgr.x (mgr.14152) 468 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:37.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:36 vm04 bash[20194]: audit 2026-03-21T06:58:36.752504+0000 mgr.x (mgr.14152) 468 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:37.900 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:58:37 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:58:38.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:37 vm07 bash[19945]: audit 2026-03-21T06:58:37.594686+0000 mgr.x (mgr.14152) 469 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:38.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:37 vm07 bash[19945]: audit 2026-03-21T06:58:37.594686+0000 mgr.x (mgr.14152) 469 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:38.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:37 vm07 bash[19945]: cluster 2026-03-21T06:58:37.632537+0000 mgr.x (mgr.14152) 470 : cluster [DBG] pgmap v373: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:38.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:37 vm07 bash[19945]: cluster 2026-03-21T06:58:37.632537+0000 mgr.x (mgr.14152) 470 : cluster [DBG] pgmap v373: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:38.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:37 vm04 bash[20194]: audit 2026-03-21T06:58:37.594686+0000 mgr.x (mgr.14152) 469 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:38.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:37 vm04 bash[20194]: audit 2026-03-21T06:58:37.594686+0000 mgr.x (mgr.14152) 469 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:38.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:37 vm04 bash[20194]: cluster 2026-03-21T06:58:37.632537+0000 mgr.x (mgr.14152) 470 : cluster [DBG] pgmap v373: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:38.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:37 vm04 bash[20194]: cluster 2026-03-21T06:58:37.632537+0000 mgr.x (mgr.14152) 470 : cluster [DBG] pgmap v373: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:38.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:37 vm02 bash[17657]: audit 2026-03-21T06:58:37.594686+0000 mgr.x (mgr.14152) 469 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:38.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:37 vm02 bash[17657]: audit 2026-03-21T06:58:37.594686+0000 mgr.x (mgr.14152) 469 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:38.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:37 vm02 bash[17657]: cluster 2026-03-21T06:58:37.632537+0000 mgr.x (mgr.14152) 470 : cluster [DBG] pgmap v373: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:38.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:37 vm02 bash[17657]: cluster 2026-03-21T06:58:37.632537+0000 mgr.x (mgr.14152) 470 : cluster [DBG] pgmap v373: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:41.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:40 vm04 bash[20194]: cluster 2026-03-21T06:58:39.632878+0000 mgr.x (mgr.14152) 471 : cluster [DBG] pgmap v374: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:41.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:40 vm04 bash[20194]: cluster 2026-03-21T06:58:39.632878+0000 mgr.x (mgr.14152) 471 : cluster [DBG] pgmap v374: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:41.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:40 vm02 bash[17657]: cluster 2026-03-21T06:58:39.632878+0000 mgr.x (mgr.14152) 471 : cluster [DBG] pgmap v374: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:41.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:40 vm02 bash[17657]: cluster 2026-03-21T06:58:39.632878+0000 mgr.x (mgr.14152) 471 : cluster [DBG] pgmap v374: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:41.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:40 vm07 bash[19945]: cluster 2026-03-21T06:58:39.632878+0000 mgr.x (mgr.14152) 471 : cluster [DBG] pgmap v374: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:41.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:40 vm07 bash[19945]: cluster 2026-03-21T06:58:39.632878+0000 mgr.x (mgr.14152) 471 : cluster [DBG] pgmap v374: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:42.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:41 vm02 bash[17657]: cluster 2026-03-21T06:58:41.633271+0000 mgr.x (mgr.14152) 472 : cluster [DBG] pgmap v375: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:42.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:41 vm02 bash[17657]: cluster 2026-03-21T06:58:41.633271+0000 mgr.x (mgr.14152) 472 : cluster [DBG] pgmap v375: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:42.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:41 vm07 bash[19945]: cluster 2026-03-21T06:58:41.633271+0000 mgr.x (mgr.14152) 472 : cluster [DBG] pgmap v375: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:42.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:41 vm07 bash[19945]: cluster 2026-03-21T06:58:41.633271+0000 mgr.x (mgr.14152) 472 : cluster [DBG] pgmap v375: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:42.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:41 vm04 bash[20194]: cluster 2026-03-21T06:58:41.633271+0000 mgr.x (mgr.14152) 472 : cluster [DBG] pgmap v375: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:42.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:41 vm04 bash[20194]: cluster 2026-03-21T06:58:41.633271+0000 mgr.x (mgr.14152) 472 : cluster [DBG] pgmap v375: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:45.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:44 vm04 bash[20194]: cluster 2026-03-21T06:58:43.633599+0000 mgr.x (mgr.14152) 473 : cluster [DBG] pgmap v376: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:45.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:44 vm04 bash[20194]: cluster 2026-03-21T06:58:43.633599+0000 mgr.x (mgr.14152) 473 : cluster [DBG] pgmap v376: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:45.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:44 vm02 bash[17657]: cluster 2026-03-21T06:58:43.633599+0000 mgr.x (mgr.14152) 473 : cluster [DBG] pgmap v376: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:45.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:44 vm02 bash[17657]: cluster 2026-03-21T06:58:43.633599+0000 mgr.x (mgr.14152) 473 : cluster [DBG] pgmap v376: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:45.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:44 vm07 bash[19945]: cluster 2026-03-21T06:58:43.633599+0000 mgr.x (mgr.14152) 473 : cluster [DBG] pgmap v376: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:45.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:44 vm07 bash[19945]: cluster 2026-03-21T06:58:43.633599+0000 mgr.x (mgr.14152) 473 : cluster [DBG] pgmap v376: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:46 vm04 bash[20194]: cluster 2026-03-21T06:58:45.633890+0000 mgr.x (mgr.14152) 474 : cluster [DBG] pgmap v377: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:46 vm04 bash[20194]: cluster 2026-03-21T06:58:45.633890+0000 mgr.x (mgr.14152) 474 : cluster [DBG] pgmap v377: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:47.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:58:46 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:58:47.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:46 vm02 bash[17657]: cluster 2026-03-21T06:58:45.633890+0000 mgr.x (mgr.14152) 474 : cluster [DBG] pgmap v377: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:47.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:46 vm02 bash[17657]: cluster 2026-03-21T06:58:45.633890+0000 mgr.x (mgr.14152) 474 : cluster [DBG] pgmap v377: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:47.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:46 vm07 bash[19945]: cluster 2026-03-21T06:58:45.633890+0000 mgr.x (mgr.14152) 474 : cluster [DBG] pgmap v377: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:47.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:46 vm07 bash[19945]: cluster 2026-03-21T06:58:45.633890+0000 mgr.x (mgr.14152) 474 : cluster [DBG] pgmap v377: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:47.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:58:47 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:58:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:47 vm07 bash[19945]: audit 2026-03-21T06:58:46.760365+0000 mgr.x (mgr.14152) 475 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:47 vm07 bash[19945]: audit 2026-03-21T06:58:46.760365+0000 mgr.x (mgr.14152) 475 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:48.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:47 vm04 bash[20194]: audit 2026-03-21T06:58:46.760365+0000 mgr.x (mgr.14152) 475 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:48.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:47 vm04 bash[20194]: audit 2026-03-21T06:58:46.760365+0000 mgr.x (mgr.14152) 475 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:48.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:47 vm02 bash[17657]: audit 2026-03-21T06:58:46.760365+0000 mgr.x (mgr.14152) 475 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:48.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:47 vm02 bash[17657]: audit 2026-03-21T06:58:46.760365+0000 mgr.x (mgr.14152) 475 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:49.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:48 vm04 bash[20194]: audit 2026-03-21T06:58:47.598399+0000 mgr.x (mgr.14152) 476 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:49.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:48 vm04 bash[20194]: audit 2026-03-21T06:58:47.598399+0000 mgr.x (mgr.14152) 476 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:49.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:48 vm04 bash[20194]: cluster 2026-03-21T06:58:47.634146+0000 mgr.x (mgr.14152) 477 : cluster [DBG] pgmap v378: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:49.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:48 vm04 bash[20194]: cluster 2026-03-21T06:58:47.634146+0000 mgr.x (mgr.14152) 477 : cluster [DBG] pgmap v378: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:48 vm02 bash[17657]: audit 2026-03-21T06:58:47.598399+0000 mgr.x (mgr.14152) 476 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:48 vm02 bash[17657]: audit 2026-03-21T06:58:47.598399+0000 mgr.x (mgr.14152) 476 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:48 vm02 bash[17657]: cluster 2026-03-21T06:58:47.634146+0000 mgr.x (mgr.14152) 477 : cluster [DBG] pgmap v378: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:48 vm02 bash[17657]: cluster 2026-03-21T06:58:47.634146+0000 mgr.x (mgr.14152) 477 : cluster [DBG] pgmap v378: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:48 vm07 bash[19945]: audit 2026-03-21T06:58:47.598399+0000 mgr.x (mgr.14152) 476 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:48 vm07 bash[19945]: audit 2026-03-21T06:58:47.598399+0000 mgr.x (mgr.14152) 476 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:48 vm07 bash[19945]: cluster 2026-03-21T06:58:47.634146+0000 mgr.x (mgr.14152) 477 : cluster [DBG] pgmap v378: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:48 vm07 bash[19945]: cluster 2026-03-21T06:58:47.634146+0000 mgr.x (mgr.14152) 477 : cluster [DBG] pgmap v378: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:51.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:50 vm02 bash[17657]: cluster 2026-03-21T06:58:49.634437+0000 mgr.x (mgr.14152) 478 : cluster [DBG] pgmap v379: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:51.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:50 vm02 bash[17657]: cluster 2026-03-21T06:58:49.634437+0000 mgr.x (mgr.14152) 478 : cluster [DBG] pgmap v379: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:50 vm07 bash[19945]: cluster 2026-03-21T06:58:49.634437+0000 mgr.x (mgr.14152) 478 : cluster [DBG] pgmap v379: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:50 vm07 bash[19945]: cluster 2026-03-21T06:58:49.634437+0000 mgr.x (mgr.14152) 478 : cluster [DBG] pgmap v379: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:50 vm04 bash[20194]: cluster 2026-03-21T06:58:49.634437+0000 mgr.x (mgr.14152) 478 : cluster [DBG] pgmap v379: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:50 vm04 bash[20194]: cluster 2026-03-21T06:58:49.634437+0000 mgr.x (mgr.14152) 478 : cluster [DBG] pgmap v379: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:51 vm02 bash[17657]: cluster 2026-03-21T06:58:51.634793+0000 mgr.x (mgr.14152) 479 : cluster [DBG] pgmap v380: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:51 vm02 bash[17657]: cluster 2026-03-21T06:58:51.634793+0000 mgr.x (mgr.14152) 479 : cluster [DBG] pgmap v380: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:52.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:51 vm07 bash[19945]: cluster 2026-03-21T06:58:51.634793+0000 mgr.x (mgr.14152) 479 : cluster [DBG] pgmap v380: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:52.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:51 vm07 bash[19945]: cluster 2026-03-21T06:58:51.634793+0000 mgr.x (mgr.14152) 479 : cluster [DBG] pgmap v380: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:52.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:51 vm04 bash[20194]: cluster 2026-03-21T06:58:51.634793+0000 mgr.x (mgr.14152) 479 : cluster [DBG] pgmap v380: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:52.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:51 vm04 bash[20194]: cluster 2026-03-21T06:58:51.634793+0000 mgr.x (mgr.14152) 479 : cluster [DBG] pgmap v380: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:55.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:54 vm04 bash[20194]: cluster 2026-03-21T06:58:53.635162+0000 mgr.x (mgr.14152) 480 : cluster [DBG] pgmap v381: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:55.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:54 vm04 bash[20194]: cluster 2026-03-21T06:58:53.635162+0000 mgr.x (mgr.14152) 480 : cluster [DBG] pgmap v381: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:55.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:54 vm02 bash[17657]: cluster 2026-03-21T06:58:53.635162+0000 mgr.x (mgr.14152) 480 : cluster [DBG] pgmap v381: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:55.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:54 vm02 bash[17657]: cluster 2026-03-21T06:58:53.635162+0000 mgr.x (mgr.14152) 480 : cluster [DBG] pgmap v381: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:55.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:54 vm07 bash[19945]: cluster 2026-03-21T06:58:53.635162+0000 mgr.x (mgr.14152) 480 : cluster [DBG] pgmap v381: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:55.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:54 vm07 bash[19945]: cluster 2026-03-21T06:58:53.635162+0000 mgr.x (mgr.14152) 480 : cluster [DBG] pgmap v381: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:56.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:56 vm07 bash[19945]: cluster 2026-03-21T06:58:55.635473+0000 mgr.x (mgr.14152) 481 : cluster [DBG] pgmap v382: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:56.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:56 vm07 bash[19945]: cluster 2026-03-21T06:58:55.635473+0000 mgr.x (mgr.14152) 481 : cluster [DBG] pgmap v382: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:56 vm04 bash[20194]: cluster 2026-03-21T06:58:55.635473+0000 mgr.x (mgr.14152) 481 : cluster [DBG] pgmap v382: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:56.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:56 vm04 bash[20194]: cluster 2026-03-21T06:58:55.635473+0000 mgr.x (mgr.14152) 481 : cluster [DBG] pgmap v382: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:56.767 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:56 vm02 bash[17657]: cluster 2026-03-21T06:58:55.635473+0000 mgr.x (mgr.14152) 481 : cluster [DBG] pgmap v382: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:56.767 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:56 vm02 bash[17657]: cluster 2026-03-21T06:58:55.635473+0000 mgr.x (mgr.14152) 481 : cluster [DBG] pgmap v382: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:58:57.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:58:56 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:58:57.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:58:57 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:58:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:58 vm04 bash[20194]: audit 2026-03-21T06:58:56.770926+0000 mgr.x (mgr.14152) 482 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:58 vm04 bash[20194]: audit 2026-03-21T06:58:56.770926+0000 mgr.x (mgr.14152) 482 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:58 vm04 bash[20194]: audit 2026-03-21T06:58:57.609261+0000 mgr.x (mgr.14152) 483 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:58 vm04 bash[20194]: audit 2026-03-21T06:58:57.609261+0000 mgr.x (mgr.14152) 483 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:58 vm04 bash[20194]: cluster 2026-03-21T06:58:57.635759+0000 mgr.x (mgr.14152) 484 : cluster [DBG] pgmap v383: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:58:58 vm04 bash[20194]: cluster 2026-03-21T06:58:57.635759+0000 mgr.x (mgr.14152) 484 : cluster [DBG] pgmap v383: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:58 vm02 bash[17657]: audit 2026-03-21T06:58:56.770926+0000 mgr.x (mgr.14152) 482 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:58 vm02 bash[17657]: audit 2026-03-21T06:58:56.770926+0000 mgr.x (mgr.14152) 482 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:58 vm02 bash[17657]: audit 2026-03-21T06:58:57.609261+0000 mgr.x (mgr.14152) 483 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:58 vm02 bash[17657]: audit 2026-03-21T06:58:57.609261+0000 mgr.x (mgr.14152) 483 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:58 vm02 bash[17657]: cluster 2026-03-21T06:58:57.635759+0000 mgr.x (mgr.14152) 484 : cluster [DBG] pgmap v383: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:58:58 vm02 bash[17657]: cluster 2026-03-21T06:58:57.635759+0000 mgr.x (mgr.14152) 484 : cluster [DBG] pgmap v383: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:58 vm07 bash[19945]: audit 2026-03-21T06:58:56.770926+0000 mgr.x (mgr.14152) 482 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:58 vm07 bash[19945]: audit 2026-03-21T06:58:56.770926+0000 mgr.x (mgr.14152) 482 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:58 vm07 bash[19945]: audit 2026-03-21T06:58:57.609261+0000 mgr.x (mgr.14152) 483 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:58 vm07 bash[19945]: audit 2026-03-21T06:58:57.609261+0000 mgr.x (mgr.14152) 483 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:58:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:58 vm07 bash[19945]: cluster 2026-03-21T06:58:57.635759+0000 mgr.x (mgr.14152) 484 : cluster [DBG] pgmap v383: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:58:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:58:58 vm07 bash[19945]: cluster 2026-03-21T06:58:57.635759+0000 mgr.x (mgr.14152) 484 : cluster [DBG] pgmap v383: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:01.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:00 vm04 bash[20194]: cluster 2026-03-21T06:58:59.636093+0000 mgr.x (mgr.14152) 485 : cluster [DBG] pgmap v384: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:01.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:00 vm04 bash[20194]: cluster 2026-03-21T06:58:59.636093+0000 mgr.x (mgr.14152) 485 : cluster [DBG] pgmap v384: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:01.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:00 vm02 bash[17657]: cluster 2026-03-21T06:58:59.636093+0000 mgr.x (mgr.14152) 485 : cluster [DBG] pgmap v384: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:01.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:00 vm02 bash[17657]: cluster 2026-03-21T06:58:59.636093+0000 mgr.x (mgr.14152) 485 : cluster [DBG] pgmap v384: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:01.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:00 vm07 bash[19945]: cluster 2026-03-21T06:58:59.636093+0000 mgr.x (mgr.14152) 485 : cluster [DBG] pgmap v384: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:01.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:00 vm07 bash[19945]: cluster 2026-03-21T06:58:59.636093+0000 mgr.x (mgr.14152) 485 : cluster [DBG] pgmap v384: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:02 vm04 bash[20194]: cluster 2026-03-21T06:59:01.636482+0000 mgr.x (mgr.14152) 486 : cluster [DBG] pgmap v385: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:03.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:02 vm04 bash[20194]: cluster 2026-03-21T06:59:01.636482+0000 mgr.x (mgr.14152) 486 : cluster [DBG] pgmap v385: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:03.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:02 vm02 bash[17657]: cluster 2026-03-21T06:59:01.636482+0000 mgr.x (mgr.14152) 486 : cluster [DBG] pgmap v385: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:03.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:02 vm02 bash[17657]: cluster 2026-03-21T06:59:01.636482+0000 mgr.x (mgr.14152) 486 : cluster [DBG] pgmap v385: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:03.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:02 vm07 bash[19945]: cluster 2026-03-21T06:59:01.636482+0000 mgr.x (mgr.14152) 486 : cluster [DBG] pgmap v385: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:03.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:02 vm07 bash[19945]: cluster 2026-03-21T06:59:01.636482+0000 mgr.x (mgr.14152) 486 : cluster [DBG] pgmap v385: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:04 vm04 bash[20194]: cluster 2026-03-21T06:59:03.636823+0000 mgr.x (mgr.14152) 487 : cluster [DBG] pgmap v386: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:05.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:04 vm04 bash[20194]: cluster 2026-03-21T06:59:03.636823+0000 mgr.x (mgr.14152) 487 : cluster [DBG] pgmap v386: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:05.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:04 vm02 bash[17657]: cluster 2026-03-21T06:59:03.636823+0000 mgr.x (mgr.14152) 487 : cluster [DBG] pgmap v386: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:05.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:04 vm02 bash[17657]: cluster 2026-03-21T06:59:03.636823+0000 mgr.x (mgr.14152) 487 : cluster [DBG] pgmap v386: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:05.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:04 vm07 bash[19945]: cluster 2026-03-21T06:59:03.636823+0000 mgr.x (mgr.14152) 487 : cluster [DBG] pgmap v386: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:05.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:04 vm07 bash[19945]: cluster 2026-03-21T06:59:03.636823+0000 mgr.x (mgr.14152) 487 : cluster [DBG] pgmap v386: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:07.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:06 vm04 bash[20194]: cluster 2026-03-21T06:59:05.637255+0000 mgr.x (mgr.14152) 488 : cluster [DBG] pgmap v387: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:07.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:06 vm04 bash[20194]: cluster 2026-03-21T06:59:05.637255+0000 mgr.x (mgr.14152) 488 : cluster [DBG] pgmap v387: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:07.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:06 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:59:07.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:06 vm02 bash[17657]: cluster 2026-03-21T06:59:05.637255+0000 mgr.x (mgr.14152) 488 : cluster [DBG] pgmap v387: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:07.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:06 vm02 bash[17657]: cluster 2026-03-21T06:59:05.637255+0000 mgr.x (mgr.14152) 488 : cluster [DBG] pgmap v387: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:07.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:06 vm07 bash[19945]: cluster 2026-03-21T06:59:05.637255+0000 mgr.x (mgr.14152) 488 : cluster [DBG] pgmap v387: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:07.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:06 vm07 bash[19945]: cluster 2026-03-21T06:59:05.637255+0000 mgr.x (mgr.14152) 488 : cluster [DBG] pgmap v387: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:07.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:07 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:59:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:08 vm04 bash[20194]: audit 2026-03-21T06:59:06.781618+0000 mgr.x (mgr.14152) 489 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:08 vm04 bash[20194]: audit 2026-03-21T06:59:06.781618+0000 mgr.x (mgr.14152) 489 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:08 vm04 bash[20194]: audit 2026-03-21T06:59:07.616480+0000 mgr.x (mgr.14152) 490 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:08 vm04 bash[20194]: audit 2026-03-21T06:59:07.616480+0000 mgr.x (mgr.14152) 490 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:08 vm04 bash[20194]: cluster 2026-03-21T06:59:07.637643+0000 mgr.x (mgr.14152) 491 : cluster [DBG] pgmap v388: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:08 vm04 bash[20194]: cluster 2026-03-21T06:59:07.637643+0000 mgr.x (mgr.14152) 491 : cluster [DBG] pgmap v388: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:09.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:08 vm02 bash[17657]: audit 2026-03-21T06:59:06.781618+0000 mgr.x (mgr.14152) 489 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:08 vm02 bash[17657]: audit 2026-03-21T06:59:06.781618+0000 mgr.x (mgr.14152) 489 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:08 vm02 bash[17657]: audit 2026-03-21T06:59:07.616480+0000 mgr.x (mgr.14152) 490 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:08 vm02 bash[17657]: audit 2026-03-21T06:59:07.616480+0000 mgr.x (mgr.14152) 490 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:08 vm02 bash[17657]: cluster 2026-03-21T06:59:07.637643+0000 mgr.x (mgr.14152) 491 : cluster [DBG] pgmap v388: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:09.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:08 vm02 bash[17657]: cluster 2026-03-21T06:59:07.637643+0000 mgr.x (mgr.14152) 491 : cluster [DBG] pgmap v388: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:08 vm07 bash[19945]: audit 2026-03-21T06:59:06.781618+0000 mgr.x (mgr.14152) 489 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:08 vm07 bash[19945]: audit 2026-03-21T06:59:06.781618+0000 mgr.x (mgr.14152) 489 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:08 vm07 bash[19945]: audit 2026-03-21T06:59:07.616480+0000 mgr.x (mgr.14152) 490 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:08 vm07 bash[19945]: audit 2026-03-21T06:59:07.616480+0000 mgr.x (mgr.14152) 490 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:08 vm07 bash[19945]: cluster 2026-03-21T06:59:07.637643+0000 mgr.x (mgr.14152) 491 : cluster [DBG] pgmap v388: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:08 vm07 bash[19945]: cluster 2026-03-21T06:59:07.637643+0000 mgr.x (mgr.14152) 491 : cluster [DBG] pgmap v388: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:11.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:10 vm02 bash[17657]: cluster 2026-03-21T06:59:09.637941+0000 mgr.x (mgr.14152) 492 : cluster [DBG] pgmap v389: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:11.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:10 vm02 bash[17657]: cluster 2026-03-21T06:59:09.637941+0000 mgr.x (mgr.14152) 492 : cluster [DBG] pgmap v389: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:10 vm07 bash[19945]: cluster 2026-03-21T06:59:09.637941+0000 mgr.x (mgr.14152) 492 : cluster [DBG] pgmap v389: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:10 vm07 bash[19945]: cluster 2026-03-21T06:59:09.637941+0000 mgr.x (mgr.14152) 492 : cluster [DBG] pgmap v389: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:11.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:10 vm04 bash[20194]: cluster 2026-03-21T06:59:09.637941+0000 mgr.x (mgr.14152) 492 : cluster [DBG] pgmap v389: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:11.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:10 vm04 bash[20194]: cluster 2026-03-21T06:59:09.637941+0000 mgr.x (mgr.14152) 492 : cluster [DBG] pgmap v389: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:12.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:11 vm02 bash[17657]: cluster 2026-03-21T06:59:11.638255+0000 mgr.x (mgr.14152) 493 : cluster [DBG] pgmap v390: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:12.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:11 vm02 bash[17657]: cluster 2026-03-21T06:59:11.638255+0000 mgr.x (mgr.14152) 493 : cluster [DBG] pgmap v390: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:12.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:11 vm07 bash[19945]: cluster 2026-03-21T06:59:11.638255+0000 mgr.x (mgr.14152) 493 : cluster [DBG] pgmap v390: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:12.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:11 vm07 bash[19945]: cluster 2026-03-21T06:59:11.638255+0000 mgr.x (mgr.14152) 493 : cluster [DBG] pgmap v390: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:12.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:11 vm04 bash[20194]: cluster 2026-03-21T06:59:11.638255+0000 mgr.x (mgr.14152) 493 : cluster [DBG] pgmap v390: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:12.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:11 vm04 bash[20194]: cluster 2026-03-21T06:59:11.638255+0000 mgr.x (mgr.14152) 493 : cluster [DBG] pgmap v390: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:15.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:14 vm04 bash[20194]: cluster 2026-03-21T06:59:13.638599+0000 mgr.x (mgr.14152) 494 : cluster [DBG] pgmap v391: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:15.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:14 vm04 bash[20194]: cluster 2026-03-21T06:59:13.638599+0000 mgr.x (mgr.14152) 494 : cluster [DBG] pgmap v391: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:15.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:14 vm02 bash[17657]: cluster 2026-03-21T06:59:13.638599+0000 mgr.x (mgr.14152) 494 : cluster [DBG] pgmap v391: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:15.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:14 vm02 bash[17657]: cluster 2026-03-21T06:59:13.638599+0000 mgr.x (mgr.14152) 494 : cluster [DBG] pgmap v391: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:15.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:14 vm07 bash[19945]: cluster 2026-03-21T06:59:13.638599+0000 mgr.x (mgr.14152) 494 : cluster [DBG] pgmap v391: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:15.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:14 vm07 bash[19945]: cluster 2026-03-21T06:59:13.638599+0000 mgr.x (mgr.14152) 494 : cluster [DBG] pgmap v391: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:16 vm04 bash[20194]: cluster 2026-03-21T06:59:15.638924+0000 mgr.x (mgr.14152) 495 : cluster [DBG] pgmap v392: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:16 vm04 bash[20194]: cluster 2026-03-21T06:59:15.638924+0000 mgr.x (mgr.14152) 495 : cluster [DBG] pgmap v392: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:17.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:16 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:59:17.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:16 vm02 bash[17657]: cluster 2026-03-21T06:59:15.638924+0000 mgr.x (mgr.14152) 495 : cluster [DBG] pgmap v392: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:17.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:16 vm02 bash[17657]: cluster 2026-03-21T06:59:15.638924+0000 mgr.x (mgr.14152) 495 : cluster [DBG] pgmap v392: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:17.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:16 vm07 bash[19945]: cluster 2026-03-21T06:59:15.638924+0000 mgr.x (mgr.14152) 495 : cluster [DBG] pgmap v392: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:17.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:16 vm07 bash[19945]: cluster 2026-03-21T06:59:15.638924+0000 mgr.x (mgr.14152) 495 : cluster [DBG] pgmap v392: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T06:59:17.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:17 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr:Note: switching to '0392f78529848ec72469e8e431875cb98d3a5fb4'. 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr:state without impacting any branches by switching back to a branch. 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr: git switch -c 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr:Or undo this operation with: 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr: git switch - 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T06:59:18.866 INFO:teuthology.orchestra.run.vm07.stderr:HEAD is now at 0392f785298 qa/tasks/keystone: restart mariadb for rocky and alma linux too 2026-03-21T06:59:18.873 DEBUG:teuthology.orchestra.run.vm07:> cp -- /home/ubuntu/cephtest/clone.client.2/src/test/cli-integration/rbd/gwcli_delete.t /home/ubuntu/cephtest/archive/cram.client.2 2026-03-21T06:59:18.876 INFO:tasks.cram:Running tests for client.0... 2026-03-21T06:59:18.876 DEBUG:teuthology.orchestra.run.vm02:> CEPH_REF=master CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.0/*.t 2026-03-21T06:59:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:18 vm04 bash[20194]: audit 2026-03-21T06:59:16.792087+0000 mgr.x (mgr.14152) 496 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:18 vm04 bash[20194]: audit 2026-03-21T06:59:16.792087+0000 mgr.x (mgr.14152) 496 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:18 vm04 bash[20194]: audit 2026-03-21T06:59:17.624462+0000 mgr.x (mgr.14152) 497 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:18 vm04 bash[20194]: audit 2026-03-21T06:59:17.624462+0000 mgr.x (mgr.14152) 497 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:18 vm04 bash[20194]: cluster 2026-03-21T06:59:17.639246+0000 mgr.x (mgr.14152) 498 : cluster [DBG] pgmap v393: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:18 vm04 bash[20194]: cluster 2026-03-21T06:59:17.639246+0000 mgr.x (mgr.14152) 498 : cluster [DBG] pgmap v393: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:18 vm02 bash[17657]: audit 2026-03-21T06:59:16.792087+0000 mgr.x (mgr.14152) 496 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:18 vm02 bash[17657]: audit 2026-03-21T06:59:16.792087+0000 mgr.x (mgr.14152) 496 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:18 vm02 bash[17657]: audit 2026-03-21T06:59:17.624462+0000 mgr.x (mgr.14152) 497 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:18 vm02 bash[17657]: audit 2026-03-21T06:59:17.624462+0000 mgr.x (mgr.14152) 497 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:18 vm02 bash[17657]: cluster 2026-03-21T06:59:17.639246+0000 mgr.x (mgr.14152) 498 : cluster [DBG] pgmap v393: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:18 vm02 bash[17657]: cluster 2026-03-21T06:59:17.639246+0000 mgr.x (mgr.14152) 498 : cluster [DBG] pgmap v393: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:19.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:18 vm07 bash[19945]: audit 2026-03-21T06:59:16.792087+0000 mgr.x (mgr.14152) 496 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:18 vm07 bash[19945]: audit 2026-03-21T06:59:16.792087+0000 mgr.x (mgr.14152) 496 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:18 vm07 bash[19945]: audit 2026-03-21T06:59:17.624462+0000 mgr.x (mgr.14152) 497 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:18 vm07 bash[19945]: audit 2026-03-21T06:59:17.624462+0000 mgr.x (mgr.14152) 497 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:19.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:18 vm07 bash[19945]: cluster 2026-03-21T06:59:17.639246+0000 mgr.x (mgr.14152) 498 : cluster [DBG] pgmap v393: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:19.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:18 vm07 bash[19945]: cluster 2026-03-21T06:59:17.639246+0000 mgr.x (mgr.14152) 498 : cluster [DBG] pgmap v393: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:19.750 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:19] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:19.750 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:19] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:19.886 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[17657]: audit 2026-03-21T06:59:19.409539+0000 mon.a (mon.0) 872 : audit [DBG] from='client.? 192.168.123.102:0/60184083' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:19.886 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[17657]: audit 2026-03-21T06:59:19.409539+0000 mon.a (mon.0) 872 : audit [DBG] from='client.? 192.168.123.102:0/60184083' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:19.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[17657]: audit 2026-03-21T06:59:19.430519+0000 mon.a (mon.0) 873 : audit [DBG] from='client.? 192.168.123.102:0/2555537384' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:19.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[17657]: audit 2026-03-21T06:59:19.430519+0000 mon.a (mon.0) 873 : audit [DBG] from='client.? 192.168.123.102:0/2555537384' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:19.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[17657]: audit 2026-03-21T06:59:19.431937+0000 mon.b (mon.2) 23 : audit [DBG] from='client.? 192.168.123.102:0/2504455500' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:19.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[17657]: audit 2026-03-21T06:59:19.431937+0000 mon.b (mon.2) 23 : audit [DBG] from='client.? 192.168.123.102:0/2504455500' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:19.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[17657]: audit 2026-03-21T06:59:19.478604+0000 mon.a (mon.0) 874 : audit [DBG] from='client.? 192.168.123.102:0/1874966201' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:19.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[17657]: audit 2026-03-21T06:59:19.478604+0000 mon.a (mon.0) 874 : audit [DBG] from='client.? 192.168.123.102:0/1874966201' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:19.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[17657]: audit 2026-03-21T06:59:19.486918+0000 mon.a (mon.0) 875 : audit [DBG] from='client.? 192.168.123.102:0/966100465' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:19.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[17657]: audit 2026-03-21T06:59:19.486918+0000 mon.a (mon.0) 875 : audit [DBG] from='client.? 192.168.123.102:0/966100465' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:20.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:19 vm04 bash[20194]: audit 2026-03-21T06:59:19.409539+0000 mon.a (mon.0) 872 : audit [DBG] from='client.? 192.168.123.102:0/60184083' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:20.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:19 vm04 bash[20194]: audit 2026-03-21T06:59:19.409539+0000 mon.a (mon.0) 872 : audit [DBG] from='client.? 192.168.123.102:0/60184083' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:20.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:19 vm04 bash[20194]: audit 2026-03-21T06:59:19.430519+0000 mon.a (mon.0) 873 : audit [DBG] from='client.? 192.168.123.102:0/2555537384' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:20.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:19 vm04 bash[20194]: audit 2026-03-21T06:59:19.430519+0000 mon.a (mon.0) 873 : audit [DBG] from='client.? 192.168.123.102:0/2555537384' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:20.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:19 vm04 bash[20194]: audit 2026-03-21T06:59:19.431937+0000 mon.b (mon.2) 23 : audit [DBG] from='client.? 192.168.123.102:0/2504455500' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:20.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:19 vm04 bash[20194]: audit 2026-03-21T06:59:19.431937+0000 mon.b (mon.2) 23 : audit [DBG] from='client.? 192.168.123.102:0/2504455500' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:20.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:19 vm04 bash[20194]: audit 2026-03-21T06:59:19.478604+0000 mon.a (mon.0) 874 : audit [DBG] from='client.? 192.168.123.102:0/1874966201' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:20.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:19 vm04 bash[20194]: audit 2026-03-21T06:59:19.478604+0000 mon.a (mon.0) 874 : audit [DBG] from='client.? 192.168.123.102:0/1874966201' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:20.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:19 vm04 bash[20194]: audit 2026-03-21T06:59:19.486918+0000 mon.a (mon.0) 875 : audit [DBG] from='client.? 192.168.123.102:0/966100465' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:20.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:19 vm04 bash[20194]: audit 2026-03-21T06:59:19.486918+0000 mon.a (mon.0) 875 : audit [DBG] from='client.? 192.168.123.102:0/966100465' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:20.055 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:19] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:20.055 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:19 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:19] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:20.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:19 vm07 bash[19945]: audit 2026-03-21T06:59:19.409539+0000 mon.a (mon.0) 872 : audit [DBG] from='client.? 192.168.123.102:0/60184083' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:20.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:19 vm07 bash[19945]: audit 2026-03-21T06:59:19.409539+0000 mon.a (mon.0) 872 : audit [DBG] from='client.? 192.168.123.102:0/60184083' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:20.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:19 vm07 bash[19945]: audit 2026-03-21T06:59:19.430519+0000 mon.a (mon.0) 873 : audit [DBG] from='client.? 192.168.123.102:0/2555537384' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:20.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:19 vm07 bash[19945]: audit 2026-03-21T06:59:19.430519+0000 mon.a (mon.0) 873 : audit [DBG] from='client.? 192.168.123.102:0/2555537384' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:20.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:19 vm07 bash[19945]: audit 2026-03-21T06:59:19.431937+0000 mon.b (mon.2) 23 : audit [DBG] from='client.? 192.168.123.102:0/2504455500' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:20.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:19 vm07 bash[19945]: audit 2026-03-21T06:59:19.431937+0000 mon.b (mon.2) 23 : audit [DBG] from='client.? 192.168.123.102:0/2504455500' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:20.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:19 vm07 bash[19945]: audit 2026-03-21T06:59:19.478604+0000 mon.a (mon.0) 874 : audit [DBG] from='client.? 192.168.123.102:0/1874966201' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:20.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:19 vm07 bash[19945]: audit 2026-03-21T06:59:19.478604+0000 mon.a (mon.0) 874 : audit [DBG] from='client.? 192.168.123.102:0/1874966201' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:20.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:19 vm07 bash[19945]: audit 2026-03-21T06:59:19.486918+0000 mon.a (mon.0) 875 : audit [DBG] from='client.? 192.168.123.102:0/966100465' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:20.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:19 vm07 bash[19945]: audit 2026-03-21T06:59:19.486918+0000 mon.a (mon.0) 875 : audit [DBG] from='client.? 192.168.123.102:0/966100465' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug (LUN.allocate) created datapool/block0 successfully 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block0' to LIO backstore user:rbd 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug (LUN.add_dev_to_lio) Successfully added datapool/block0 to LIO 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug (LUN.allocate) added 'datapool/block0' to LIO and config object 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "PUT /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "PUT /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug _disk update on localhost, successful 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "PUT /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "PUT /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: cluster 2026-03-21T06:59:19.639575+0000 mgr.x (mgr.14152) 499 : cluster [DBG] pgmap v394: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: cluster 2026-03-21T06:59:19.639575+0000 mgr.x (mgr.14152) 499 : cluster [DBG] pgmap v394: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:19.932999+0000 mon.c (mon.1) 36 : audit [DBG] from='client.? 192.168.123.102:0/2331780756' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:19.932999+0000 mon.c (mon.1) 36 : audit [DBG] from='client.? 192.168.123.102:0/2331780756' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:19.963671+0000 mon.a (mon.0) 876 : audit [DBG] from='client.? 192.168.123.102:0/42282083' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:19.963671+0000 mon.a (mon.0) 876 : audit [DBG] from='client.? 192.168.123.102:0/42282083' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:19.976358+0000 mon.a (mon.0) 877 : audit [DBG] from='client.? 192.168.123.102:0/3176464274' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:19.976358+0000 mon.a (mon.0) 877 : audit [DBG] from='client.? 192.168.123.102:0/3176464274' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.015443+0000 mon.c (mon.1) 37 : audit [DBG] from='client.? 192.168.123.102:0/2148850333' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.015443+0000 mon.c (mon.1) 37 : audit [DBG] from='client.? 192.168.123.102:0/2148850333' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.023715+0000 mon.b (mon.2) 24 : audit [DBG] from='client.? 192.168.123.102:0/3860420622' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.023715+0000 mon.b (mon.2) 24 : audit [DBG] from='client.? 192.168.123.102:0/3860420622' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.161588+0000 mon.a (mon.0) 878 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.161588+0000 mon.a (mon.0) 878 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.162553+0000 mon.a (mon.0) 879 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.162553+0000 mon.a (mon.0) 879 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.162820+0000 mon.a (mon.0) 880 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.162820+0000 mon.a (mon.0) 880 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.230810+0000 mon.a (mon.0) 881 : audit [DBG] from='client.? 192.168.123.102:0/4252502130' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.230810+0000 mon.a (mon.0) 881 : audit [DBG] from='client.? 192.168.123.102:0/4252502130' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.662625+0000 mon.a (mon.0) 882 : audit [DBG] from='client.? 192.168.123.102:0/3612924530' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.662625+0000 mon.a (mon.0) 882 : audit [DBG] from='client.? 192.168.123.102:0/3612924530' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.687659+0000 mon.a (mon.0) 883 : audit [DBG] from='client.? 192.168.123.102:0/2344312844' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.687659+0000 mon.a (mon.0) 883 : audit [DBG] from='client.? 192.168.123.102:0/2344312844' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.692431+0000 mon.c (mon.1) 38 : audit [DBG] from='client.? 192.168.123.102:0/1162820058' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:20 vm04 bash[20194]: audit 2026-03-21T06:59:20.692431+0000 mon.c (mon.1) 38 : audit [DBG] from='client.? 192.168.123.102:0/1162820058' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:21.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:21.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:21.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:21.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:21.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:20] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: cluster 2026-03-21T06:59:19.639575+0000 mgr.x (mgr.14152) 499 : cluster [DBG] pgmap v394: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: cluster 2026-03-21T06:59:19.639575+0000 mgr.x (mgr.14152) 499 : cluster [DBG] pgmap v394: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:19.932999+0000 mon.c (mon.1) 36 : audit [DBG] from='client.? 192.168.123.102:0/2331780756' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:19.932999+0000 mon.c (mon.1) 36 : audit [DBG] from='client.? 192.168.123.102:0/2331780756' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:19.963671+0000 mon.a (mon.0) 876 : audit [DBG] from='client.? 192.168.123.102:0/42282083' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:19.963671+0000 mon.a (mon.0) 876 : audit [DBG] from='client.? 192.168.123.102:0/42282083' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:19.976358+0000 mon.a (mon.0) 877 : audit [DBG] from='client.? 192.168.123.102:0/3176464274' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:19.976358+0000 mon.a (mon.0) 877 : audit [DBG] from='client.? 192.168.123.102:0/3176464274' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.015443+0000 mon.c (mon.1) 37 : audit [DBG] from='client.? 192.168.123.102:0/2148850333' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.015443+0000 mon.c (mon.1) 37 : audit [DBG] from='client.? 192.168.123.102:0/2148850333' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.023715+0000 mon.b (mon.2) 24 : audit [DBG] from='client.? 192.168.123.102:0/3860420622' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.023715+0000 mon.b (mon.2) 24 : audit [DBG] from='client.? 192.168.123.102:0/3860420622' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.161588+0000 mon.a (mon.0) 878 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.161588+0000 mon.a (mon.0) 878 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.162553+0000 mon.a (mon.0) 879 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.162553+0000 mon.a (mon.0) 879 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.162820+0000 mon.a (mon.0) 880 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.162820+0000 mon.a (mon.0) 880 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.230810+0000 mon.a (mon.0) 881 : audit [DBG] from='client.? 192.168.123.102:0/4252502130' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.230810+0000 mon.a (mon.0) 881 : audit [DBG] from='client.? 192.168.123.102:0/4252502130' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.662625+0000 mon.a (mon.0) 882 : audit [DBG] from='client.? 192.168.123.102:0/3612924530' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.662625+0000 mon.a (mon.0) 882 : audit [DBG] from='client.? 192.168.123.102:0/3612924530' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.687659+0000 mon.a (mon.0) 883 : audit [DBG] from='client.? 192.168.123.102:0/2344312844' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.687659+0000 mon.a (mon.0) 883 : audit [DBG] from='client.? 192.168.123.102:0/2344312844' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.692431+0000 mon.c (mon.1) 38 : audit [DBG] from='client.? 192.168.123.102:0/1162820058' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:20 vm02 bash[17657]: audit 2026-03-21T06:59:20.692431+0000 mon.c (mon.1) 38 : audit [DBG] from='client.? 192.168.123.102:0/1162820058' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:20 vm07 bash[42142]: debug Change detected - internal 0 / xattr 1 refreshing 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: cluster 2026-03-21T06:59:19.639575+0000 mgr.x (mgr.14152) 499 : cluster [DBG] pgmap v394: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: cluster 2026-03-21T06:59:19.639575+0000 mgr.x (mgr.14152) 499 : cluster [DBG] pgmap v394: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:19.932999+0000 mon.c (mon.1) 36 : audit [DBG] from='client.? 192.168.123.102:0/2331780756' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:19.932999+0000 mon.c (mon.1) 36 : audit [DBG] from='client.? 192.168.123.102:0/2331780756' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:19.963671+0000 mon.a (mon.0) 876 : audit [DBG] from='client.? 192.168.123.102:0/42282083' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:19.963671+0000 mon.a (mon.0) 876 : audit [DBG] from='client.? 192.168.123.102:0/42282083' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:19.976358+0000 mon.a (mon.0) 877 : audit [DBG] from='client.? 192.168.123.102:0/3176464274' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:19.976358+0000 mon.a (mon.0) 877 : audit [DBG] from='client.? 192.168.123.102:0/3176464274' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.015443+0000 mon.c (mon.1) 37 : audit [DBG] from='client.? 192.168.123.102:0/2148850333' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.015443+0000 mon.c (mon.1) 37 : audit [DBG] from='client.? 192.168.123.102:0/2148850333' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.023715+0000 mon.b (mon.2) 24 : audit [DBG] from='client.? 192.168.123.102:0/3860420622' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.023715+0000 mon.b (mon.2) 24 : audit [DBG] from='client.? 192.168.123.102:0/3860420622' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.161588+0000 mon.a (mon.0) 878 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.161588+0000 mon.a (mon.0) 878 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.162553+0000 mon.a (mon.0) 879 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.162553+0000 mon.a (mon.0) 879 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.162820+0000 mon.a (mon.0) 880 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.162820+0000 mon.a (mon.0) 880 : audit [DBG] from='client.? 192.168.123.102:0/2640075721' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.230810+0000 mon.a (mon.0) 881 : audit [DBG] from='client.? 192.168.123.102:0/4252502130' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.230810+0000 mon.a (mon.0) 881 : audit [DBG] from='client.? 192.168.123.102:0/4252502130' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.662625+0000 mon.a (mon.0) 882 : audit [DBG] from='client.? 192.168.123.102:0/3612924530' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.662625+0000 mon.a (mon.0) 882 : audit [DBG] from='client.? 192.168.123.102:0/3612924530' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.687659+0000 mon.a (mon.0) 883 : audit [DBG] from='client.? 192.168.123.102:0/2344312844' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.687659+0000 mon.a (mon.0) 883 : audit [DBG] from='client.? 192.168.123.102:0/2344312844' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.692431+0000 mon.c (mon.1) 38 : audit [DBG] from='client.? 192.168.123.102:0/1162820058' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:20 vm07 bash[19945]: audit 2026-03-21T06:59:20.692431+0000 mon.c (mon.1) 38 : audit [DBG] from='client.? 192.168.123.102:0/1162820058' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:21.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:21.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:21.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:21.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:21.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:21.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:22.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:22.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:22.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:22.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:22.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:22.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:21] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:20.781950+0000 mon.a (mon.0) 884 : audit [DBG] from='client.? 192.168.123.102:0/3294568523' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:20.781950+0000 mon.a (mon.0) 884 : audit [DBG] from='client.? 192.168.123.102:0/3294568523' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:20.790930+0000 mon.a (mon.0) 885 : audit [DBG] from='client.? 192.168.123.102:0/3292256935' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:20.790930+0000 mon.a (mon.0) 885 : audit [DBG] from='client.? 192.168.123.102:0/3292256935' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.240595+0000 mon.a (mon.0) 886 : audit [DBG] from='client.? 192.168.123.102:0/1768976448' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.240595+0000 mon.a (mon.0) 886 : audit [DBG] from='client.? 192.168.123.102:0/1768976448' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.264898+0000 mon.a (mon.0) 887 : audit [DBG] from='client.? 192.168.123.102:0/2316096474' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.264898+0000 mon.a (mon.0) 887 : audit [DBG] from='client.? 192.168.123.102:0/2316096474' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.274532+0000 mon.a (mon.0) 888 : audit [DBG] from='client.? 192.168.123.102:0/194656392' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.274532+0000 mon.a (mon.0) 888 : audit [DBG] from='client.? 192.168.123.102:0/194656392' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.353999+0000 mon.a (mon.0) 889 : audit [DBG] from='client.? 192.168.123.102:0/4066571407' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.353999+0000 mon.a (mon.0) 889 : audit [DBG] from='client.? 192.168.123.102:0/4066571407' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.357776+0000 mon.c (mon.1) 39 : audit [DBG] from='client.? 192.168.123.102:0/295085168' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.357776+0000 mon.c (mon.1) 39 : audit [DBG] from='client.? 192.168.123.102:0/295085168' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.693887+0000 mon.a (mon.0) 890 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:59:22.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:21 vm02 bash[17657]: audit 2026-03-21T06:59:21.693887+0000 mon.a (mon.0) 890 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:59:22.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:20.781950+0000 mon.a (mon.0) 884 : audit [DBG] from='client.? 192.168.123.102:0/3294568523' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:20.781950+0000 mon.a (mon.0) 884 : audit [DBG] from='client.? 192.168.123.102:0/3294568523' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:20.790930+0000 mon.a (mon.0) 885 : audit [DBG] from='client.? 192.168.123.102:0/3292256935' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:20.790930+0000 mon.a (mon.0) 885 : audit [DBG] from='client.? 192.168.123.102:0/3292256935' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.240595+0000 mon.a (mon.0) 886 : audit [DBG] from='client.? 192.168.123.102:0/1768976448' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.240595+0000 mon.a (mon.0) 886 : audit [DBG] from='client.? 192.168.123.102:0/1768976448' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.264898+0000 mon.a (mon.0) 887 : audit [DBG] from='client.? 192.168.123.102:0/2316096474' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.264898+0000 mon.a (mon.0) 887 : audit [DBG] from='client.? 192.168.123.102:0/2316096474' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.274532+0000 mon.a (mon.0) 888 : audit [DBG] from='client.? 192.168.123.102:0/194656392' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.274532+0000 mon.a (mon.0) 888 : audit [DBG] from='client.? 192.168.123.102:0/194656392' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.353999+0000 mon.a (mon.0) 889 : audit [DBG] from='client.? 192.168.123.102:0/4066571407' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.353999+0000 mon.a (mon.0) 889 : audit [DBG] from='client.? 192.168.123.102:0/4066571407' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.357776+0000 mon.c (mon.1) 39 : audit [DBG] from='client.? 192.168.123.102:0/295085168' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.357776+0000 mon.c (mon.1) 39 : audit [DBG] from='client.? 192.168.123.102:0/295085168' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.693887+0000 mon.a (mon.0) 890 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:59:22.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:21 vm07 bash[19945]: audit 2026-03-21T06:59:21.693887+0000 mon.a (mon.0) 890 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:20.781950+0000 mon.a (mon.0) 884 : audit [DBG] from='client.? 192.168.123.102:0/3294568523' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:20.781950+0000 mon.a (mon.0) 884 : audit [DBG] from='client.? 192.168.123.102:0/3294568523' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:20.790930+0000 mon.a (mon.0) 885 : audit [DBG] from='client.? 192.168.123.102:0/3292256935' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:20.790930+0000 mon.a (mon.0) 885 : audit [DBG] from='client.? 192.168.123.102:0/3292256935' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.240595+0000 mon.a (mon.0) 886 : audit [DBG] from='client.? 192.168.123.102:0/1768976448' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.240595+0000 mon.a (mon.0) 886 : audit [DBG] from='client.? 192.168.123.102:0/1768976448' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.264898+0000 mon.a (mon.0) 887 : audit [DBG] from='client.? 192.168.123.102:0/2316096474' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.264898+0000 mon.a (mon.0) 887 : audit [DBG] from='client.? 192.168.123.102:0/2316096474' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.274532+0000 mon.a (mon.0) 888 : audit [DBG] from='client.? 192.168.123.102:0/194656392' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.274532+0000 mon.a (mon.0) 888 : audit [DBG] from='client.? 192.168.123.102:0/194656392' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.353999+0000 mon.a (mon.0) 889 : audit [DBG] from='client.? 192.168.123.102:0/4066571407' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.353999+0000 mon.a (mon.0) 889 : audit [DBG] from='client.? 192.168.123.102:0/4066571407' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.357776+0000 mon.c (mon.1) 39 : audit [DBG] from='client.? 192.168.123.102:0/295085168' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.357776+0000 mon.c (mon.1) 39 : audit [DBG] from='client.? 192.168.123.102:0/295085168' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.693887+0000 mon.a (mon.0) 890 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:59:22.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:21 vm04 bash[20194]: audit 2026-03-21T06:59:21.693887+0000 mon.a (mon.0) 890 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T06:59:22.771 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:22] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:22.771 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:22] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:22.771 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:22] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:22.771 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:22] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:22.771 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[49158]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T06:59:22.771 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:22] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:22.771 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:22] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:22.771 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:22] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:22.771 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:22] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:22.771 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:22] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:22.771 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:22] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: cluster 2026-03-21T06:59:21.639940+0000 mgr.x (mgr.14152) 500 : cluster [DBG] pgmap v395: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 4.1 KiB/s rd, 255 B/s wr, 5 op/s 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: cluster 2026-03-21T06:59:21.639940+0000 mgr.x (mgr.14152) 500 : cluster [DBG] pgmap v395: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 4.1 KiB/s rd, 255 B/s wr, 5 op/s 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:21.821088+0000 mon.c (mon.1) 40 : audit [DBG] from='client.? 192.168.123.102:0/1872243705' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:21.821088+0000 mon.c (mon.1) 40 : audit [DBG] from='client.? 192.168.123.102:0/1872243705' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:21.854111+0000 mon.a (mon.0) 891 : audit [DBG] from='client.? 192.168.123.102:0/961819399' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:21.854111+0000 mon.a (mon.0) 891 : audit [DBG] from='client.? 192.168.123.102:0/961819399' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:21.863892+0000 mon.a (mon.0) 892 : audit [DBG] from='client.? 192.168.123.102:0/1739506892' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:21.863892+0000 mon.a (mon.0) 892 : audit [DBG] from='client.? 192.168.123.102:0/1739506892' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:21.943923+0000 mon.a (mon.0) 893 : audit [DBG] from='client.? 192.168.123.102:0/3775306426' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:21.943923+0000 mon.a (mon.0) 893 : audit [DBG] from='client.? 192.168.123.102:0/3775306426' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:21.956815+0000 mon.a (mon.0) 894 : audit [DBG] from='client.? 192.168.123.102:0/1942284789' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:21.956815+0000 mon.a (mon.0) 894 : audit [DBG] from='client.? 192.168.123.102:0/1942284789' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.037969+0000 mon.a (mon.0) 895 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.037969+0000 mon.a (mon.0) 895 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.046888+0000 mon.a (mon.0) 896 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.046888+0000 mon.a (mon.0) 896 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.055338+0000 mon.a (mon.0) 897 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.055338+0000 mon.a (mon.0) 897 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.061937+0000 mon.a (mon.0) 898 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.061937+0000 mon.a (mon.0) 898 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.444656+0000 mon.a (mon.0) 899 : audit [DBG] from='client.? 192.168.123.102:0/3738182424' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.444656+0000 mon.a (mon.0) 899 : audit [DBG] from='client.? 192.168.123.102:0/3738182424' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.478977+0000 mon.a (mon.0) 900 : audit [DBG] from='client.? 192.168.123.102:0/2357635971' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.478977+0000 mon.a (mon.0) 900 : audit [DBG] from='client.? 192.168.123.102:0/2357635971' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.488936+0000 mon.a (mon.0) 901 : audit [DBG] from='client.? 192.168.123.102:0/3134136773' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.488936+0000 mon.a (mon.0) 901 : audit [DBG] from='client.? 192.168.123.102:0/3134136773' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.575178+0000 mon.a (mon.0) 902 : audit [DBG] from='client.? 192.168.123.102:0/1370475402' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.117 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.575178+0000 mon.a (mon.0) 902 : audit [DBG] from='client.? 192.168.123.102:0/1370475402' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.117 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.583653+0000 mon.a (mon.0) 903 : audit [DBG] from='client.? 192.168.123.102:0/1002153645' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.117 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:22 vm02 bash[17657]: audit 2026-03-21T06:59:22.583653+0000 mon.a (mon.0) 903 : audit [DBG] from='client.? 192.168.123.102:0/1002153645' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:22 vm07 bash[42142]: debug Change detected - internal 1 / xattr 2 refreshing 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: cluster 2026-03-21T06:59:21.639940+0000 mgr.x (mgr.14152) 500 : cluster [DBG] pgmap v395: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 4.1 KiB/s rd, 255 B/s wr, 5 op/s 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: cluster 2026-03-21T06:59:21.639940+0000 mgr.x (mgr.14152) 500 : cluster [DBG] pgmap v395: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 4.1 KiB/s rd, 255 B/s wr, 5 op/s 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:21.821088+0000 mon.c (mon.1) 40 : audit [DBG] from='client.? 192.168.123.102:0/1872243705' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:21.821088+0000 mon.c (mon.1) 40 : audit [DBG] from='client.? 192.168.123.102:0/1872243705' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:21.854111+0000 mon.a (mon.0) 891 : audit [DBG] from='client.? 192.168.123.102:0/961819399' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:21.854111+0000 mon.a (mon.0) 891 : audit [DBG] from='client.? 192.168.123.102:0/961819399' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:21.863892+0000 mon.a (mon.0) 892 : audit [DBG] from='client.? 192.168.123.102:0/1739506892' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:21.863892+0000 mon.a (mon.0) 892 : audit [DBG] from='client.? 192.168.123.102:0/1739506892' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:21.943923+0000 mon.a (mon.0) 893 : audit [DBG] from='client.? 192.168.123.102:0/3775306426' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:21.943923+0000 mon.a (mon.0) 893 : audit [DBG] from='client.? 192.168.123.102:0/3775306426' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:21.956815+0000 mon.a (mon.0) 894 : audit [DBG] from='client.? 192.168.123.102:0/1942284789' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:21.956815+0000 mon.a (mon.0) 894 : audit [DBG] from='client.? 192.168.123.102:0/1942284789' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.037969+0000 mon.a (mon.0) 895 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.037969+0000 mon.a (mon.0) 895 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.046888+0000 mon.a (mon.0) 896 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.046888+0000 mon.a (mon.0) 896 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.055338+0000 mon.a (mon.0) 897 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.055338+0000 mon.a (mon.0) 897 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.061937+0000 mon.a (mon.0) 898 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.061937+0000 mon.a (mon.0) 898 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.444656+0000 mon.a (mon.0) 899 : audit [DBG] from='client.? 192.168.123.102:0/3738182424' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.444656+0000 mon.a (mon.0) 899 : audit [DBG] from='client.? 192.168.123.102:0/3738182424' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.478977+0000 mon.a (mon.0) 900 : audit [DBG] from='client.? 192.168.123.102:0/2357635971' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.478977+0000 mon.a (mon.0) 900 : audit [DBG] from='client.? 192.168.123.102:0/2357635971' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.488936+0000 mon.a (mon.0) 901 : audit [DBG] from='client.? 192.168.123.102:0/3134136773' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.488936+0000 mon.a (mon.0) 901 : audit [DBG] from='client.? 192.168.123.102:0/3134136773' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.575178+0000 mon.a (mon.0) 902 : audit [DBG] from='client.? 192.168.123.102:0/1370475402' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.575178+0000 mon.a (mon.0) 902 : audit [DBG] from='client.? 192.168.123.102:0/1370475402' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.583653+0000 mon.a (mon.0) 903 : audit [DBG] from='client.? 192.168.123.102:0/1002153645' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:22 vm07 bash[19945]: audit 2026-03-21T06:59:22.583653+0000 mon.a (mon.0) 903 : audit [DBG] from='client.? 192.168.123.102:0/1002153645' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: cluster 2026-03-21T06:59:21.639940+0000 mgr.x (mgr.14152) 500 : cluster [DBG] pgmap v395: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 4.1 KiB/s rd, 255 B/s wr, 5 op/s 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: cluster 2026-03-21T06:59:21.639940+0000 mgr.x (mgr.14152) 500 : cluster [DBG] pgmap v395: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 4.1 KiB/s rd, 255 B/s wr, 5 op/s 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:21.821088+0000 mon.c (mon.1) 40 : audit [DBG] from='client.? 192.168.123.102:0/1872243705' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:21.821088+0000 mon.c (mon.1) 40 : audit [DBG] from='client.? 192.168.123.102:0/1872243705' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:21.854111+0000 mon.a (mon.0) 891 : audit [DBG] from='client.? 192.168.123.102:0/961819399' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:21.854111+0000 mon.a (mon.0) 891 : audit [DBG] from='client.? 192.168.123.102:0/961819399' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:21.863892+0000 mon.a (mon.0) 892 : audit [DBG] from='client.? 192.168.123.102:0/1739506892' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:21.863892+0000 mon.a (mon.0) 892 : audit [DBG] from='client.? 192.168.123.102:0/1739506892' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:21.943923+0000 mon.a (mon.0) 893 : audit [DBG] from='client.? 192.168.123.102:0/3775306426' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:21.943923+0000 mon.a (mon.0) 893 : audit [DBG] from='client.? 192.168.123.102:0/3775306426' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:21.956815+0000 mon.a (mon.0) 894 : audit [DBG] from='client.? 192.168.123.102:0/1942284789' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:21.956815+0000 mon.a (mon.0) 894 : audit [DBG] from='client.? 192.168.123.102:0/1942284789' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.037969+0000 mon.a (mon.0) 895 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.037969+0000 mon.a (mon.0) 895 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.046888+0000 mon.a (mon.0) 896 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.046888+0000 mon.a (mon.0) 896 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.055338+0000 mon.a (mon.0) 897 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.055338+0000 mon.a (mon.0) 897 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.061937+0000 mon.a (mon.0) 898 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.061937+0000 mon.a (mon.0) 898 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:23.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.444656+0000 mon.a (mon.0) 899 : audit [DBG] from='client.? 192.168.123.102:0/3738182424' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.444656+0000 mon.a (mon.0) 899 : audit [DBG] from='client.? 192.168.123.102:0/3738182424' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:23.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.478977+0000 mon.a (mon.0) 900 : audit [DBG] from='client.? 192.168.123.102:0/2357635971' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.478977+0000 mon.a (mon.0) 900 : audit [DBG] from='client.? 192.168.123.102:0/2357635971' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:23.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.488936+0000 mon.a (mon.0) 901 : audit [DBG] from='client.? 192.168.123.102:0/3134136773' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.488936+0000 mon.a (mon.0) 901 : audit [DBG] from='client.? 192.168.123.102:0/3134136773' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.575178+0000 mon.a (mon.0) 902 : audit [DBG] from='client.? 192.168.123.102:0/1370475402' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.575178+0000 mon.a (mon.0) 902 : audit [DBG] from='client.? 192.168.123.102:0/1370475402' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:23.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.583653+0000 mon.a (mon.0) 903 : audit [DBG] from='client.? 192.168.123.102:0/1002153645' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:22 vm04 bash[20194]: audit 2026-03-21T06:59:22.583653+0000 mon.a (mon.0) 903 : audit [DBG] from='client.? 192.168.123.102:0/1002153645' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:23.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:23.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:23.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:23.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:23.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:23.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:23.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:23.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:24.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:24.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:23] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.081436+0000 mon.a (mon.0) 904 : audit [DBG] from='client.? 192.168.123.102:0/4089220267' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.081436+0000 mon.a (mon.0) 904 : audit [DBG] from='client.? 192.168.123.102:0/4089220267' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.106145+0000 mon.a (mon.0) 905 : audit [DBG] from='client.? 192.168.123.102:0/1638003757' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.106145+0000 mon.a (mon.0) 905 : audit [DBG] from='client.? 192.168.123.102:0/1638003757' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.114817+0000 mon.a (mon.0) 906 : audit [DBG] from='client.? 192.168.123.102:0/3470582351' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.114817+0000 mon.a (mon.0) 906 : audit [DBG] from='client.? 192.168.123.102:0/3470582351' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.208424+0000 mon.a (mon.0) 907 : audit [DBG] from='client.? 192.168.123.102:0/783402312' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.208424+0000 mon.a (mon.0) 907 : audit [DBG] from='client.? 192.168.123.102:0/783402312' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.217149+0000 mon.a (mon.0) 908 : audit [DBG] from='client.? 192.168.123.102:0/799158205' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.217149+0000 mon.a (mon.0) 908 : audit [DBG] from='client.? 192.168.123.102:0/799158205' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: cluster 2026-03-21T06:59:23.640345+0000 mgr.x (mgr.14152) 501 : cluster [DBG] pgmap v396: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 3.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: cluster 2026-03-21T06:59:23.640345+0000 mgr.x (mgr.14152) 501 : cluster [DBG] pgmap v396: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 3.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.651312+0000 mon.a (mon.0) 909 : audit [DBG] from='client.? 192.168.123.102:0/2996312414' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.651312+0000 mon.a (mon.0) 909 : audit [DBG] from='client.? 192.168.123.102:0/2996312414' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.681271+0000 mon.a (mon.0) 910 : audit [DBG] from='client.? 192.168.123.102:0/3694499722' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.681271+0000 mon.a (mon.0) 910 : audit [DBG] from='client.? 192.168.123.102:0/3694499722' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.690585+0000 mon.a (mon.0) 911 : audit [DBG] from='client.? 192.168.123.102:0/3355748838' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:23 vm02 bash[17657]: audit 2026-03-21T06:59:23.690585+0000 mon.a (mon.0) 911 : audit [DBG] from='client.? 192.168.123.102:0/3355748838' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.081436+0000 mon.a (mon.0) 904 : audit [DBG] from='client.? 192.168.123.102:0/4089220267' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.081436+0000 mon.a (mon.0) 904 : audit [DBG] from='client.? 192.168.123.102:0/4089220267' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.106145+0000 mon.a (mon.0) 905 : audit [DBG] from='client.? 192.168.123.102:0/1638003757' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.106145+0000 mon.a (mon.0) 905 : audit [DBG] from='client.? 192.168.123.102:0/1638003757' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.114817+0000 mon.a (mon.0) 906 : audit [DBG] from='client.? 192.168.123.102:0/3470582351' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.114817+0000 mon.a (mon.0) 906 : audit [DBG] from='client.? 192.168.123.102:0/3470582351' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.208424+0000 mon.a (mon.0) 907 : audit [DBG] from='client.? 192.168.123.102:0/783402312' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.208424+0000 mon.a (mon.0) 907 : audit [DBG] from='client.? 192.168.123.102:0/783402312' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.217149+0000 mon.a (mon.0) 908 : audit [DBG] from='client.? 192.168.123.102:0/799158205' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.217149+0000 mon.a (mon.0) 908 : audit [DBG] from='client.? 192.168.123.102:0/799158205' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: cluster 2026-03-21T06:59:23.640345+0000 mgr.x (mgr.14152) 501 : cluster [DBG] pgmap v396: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 3.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: cluster 2026-03-21T06:59:23.640345+0000 mgr.x (mgr.14152) 501 : cluster [DBG] pgmap v396: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 3.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.651312+0000 mon.a (mon.0) 909 : audit [DBG] from='client.? 192.168.123.102:0/2996312414' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.651312+0000 mon.a (mon.0) 909 : audit [DBG] from='client.? 192.168.123.102:0/2996312414' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.681271+0000 mon.a (mon.0) 910 : audit [DBG] from='client.? 192.168.123.102:0/3694499722' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.681271+0000 mon.a (mon.0) 910 : audit [DBG] from='client.? 192.168.123.102:0/3694499722' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.690585+0000 mon.a (mon.0) 911 : audit [DBG] from='client.? 192.168.123.102:0/3355748838' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:23 vm07 bash[19945]: audit 2026-03-21T06:59:23.690585+0000 mon.a (mon.0) 911 : audit [DBG] from='client.? 192.168.123.102:0/3355748838' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.081436+0000 mon.a (mon.0) 904 : audit [DBG] from='client.? 192.168.123.102:0/4089220267' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.081436+0000 mon.a (mon.0) 904 : audit [DBG] from='client.? 192.168.123.102:0/4089220267' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.106145+0000 mon.a (mon.0) 905 : audit [DBG] from='client.? 192.168.123.102:0/1638003757' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.106145+0000 mon.a (mon.0) 905 : audit [DBG] from='client.? 192.168.123.102:0/1638003757' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.114817+0000 mon.a (mon.0) 906 : audit [DBG] from='client.? 192.168.123.102:0/3470582351' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.114817+0000 mon.a (mon.0) 906 : audit [DBG] from='client.? 192.168.123.102:0/3470582351' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.208424+0000 mon.a (mon.0) 907 : audit [DBG] from='client.? 192.168.123.102:0/783402312' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.208424+0000 mon.a (mon.0) 907 : audit [DBG] from='client.? 192.168.123.102:0/783402312' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.217149+0000 mon.a (mon.0) 908 : audit [DBG] from='client.? 192.168.123.102:0/799158205' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.217149+0000 mon.a (mon.0) 908 : audit [DBG] from='client.? 192.168.123.102:0/799158205' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: cluster 2026-03-21T06:59:23.640345+0000 mgr.x (mgr.14152) 501 : cluster [DBG] pgmap v396: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 3.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: cluster 2026-03-21T06:59:23.640345+0000 mgr.x (mgr.14152) 501 : cluster [DBG] pgmap v396: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 3.2 KiB/s rd, 255 B/s wr, 4 op/s 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.651312+0000 mon.a (mon.0) 909 : audit [DBG] from='client.? 192.168.123.102:0/2996312414' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.651312+0000 mon.a (mon.0) 909 : audit [DBG] from='client.? 192.168.123.102:0/2996312414' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.681271+0000 mon.a (mon.0) 910 : audit [DBG] from='client.? 192.168.123.102:0/3694499722' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.681271+0000 mon.a (mon.0) 910 : audit [DBG] from='client.? 192.168.123.102:0/3694499722' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.690585+0000 mon.a (mon.0) 911 : audit [DBG] from='client.? 192.168.123.102:0/3355748838' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:23 vm04 bash[20194]: audit 2026-03-21T06:59:23.690585+0000 mon.a (mon.0) 911 : audit [DBG] from='client.? 192.168.123.102:0/3355748838' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:24.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:24.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:24.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:24.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:25.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:25.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:24] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:23.792826+0000 mon.a (mon.0) 912 : audit [DBG] from='client.? 192.168.123.102:0/1991895665' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:23.792826+0000 mon.a (mon.0) 912 : audit [DBG] from='client.? 192.168.123.102:0/1991895665' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:23.803170+0000 mon.a (mon.0) 913 : audit [DBG] from='client.? 192.168.123.102:0/3010713827' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:23.803170+0000 mon.a (mon.0) 913 : audit [DBG] from='client.? 192.168.123.102:0/3010713827' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:24.254491+0000 mon.a (mon.0) 914 : audit [DBG] from='client.? 192.168.123.102:0/1281175297' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:24.254491+0000 mon.a (mon.0) 914 : audit [DBG] from='client.? 192.168.123.102:0/1281175297' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:24.282074+0000 mon.a (mon.0) 915 : audit [DBG] from='client.? 192.168.123.102:0/699741649' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:24.282074+0000 mon.a (mon.0) 915 : audit [DBG] from='client.? 192.168.123.102:0/699741649' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:24.291194+0000 mon.a (mon.0) 916 : audit [DBG] from='client.? 192.168.123.102:0/3109639298' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:24.291194+0000 mon.a (mon.0) 916 : audit [DBG] from='client.? 192.168.123.102:0/3109639298' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:24.385821+0000 mon.a (mon.0) 917 : audit [DBG] from='client.? 192.168.123.102:0/1053561522' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:24.385821+0000 mon.a (mon.0) 917 : audit [DBG] from='client.? 192.168.123.102:0/1053561522' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:24.394158+0000 mon.a (mon.0) 918 : audit [DBG] from='client.? 192.168.123.102:0/2833210967' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:24 vm02 bash[17657]: audit 2026-03-21T06:59:24.394158+0000 mon.a (mon.0) 918 : audit [DBG] from='client.? 192.168.123.102:0/2833210967' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:23.792826+0000 mon.a (mon.0) 912 : audit [DBG] from='client.? 192.168.123.102:0/1991895665' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:23.792826+0000 mon.a (mon.0) 912 : audit [DBG] from='client.? 192.168.123.102:0/1991895665' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:23.803170+0000 mon.a (mon.0) 913 : audit [DBG] from='client.? 192.168.123.102:0/3010713827' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:23.803170+0000 mon.a (mon.0) 913 : audit [DBG] from='client.? 192.168.123.102:0/3010713827' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:24.254491+0000 mon.a (mon.0) 914 : audit [DBG] from='client.? 192.168.123.102:0/1281175297' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:24.254491+0000 mon.a (mon.0) 914 : audit [DBG] from='client.? 192.168.123.102:0/1281175297' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:24.282074+0000 mon.a (mon.0) 915 : audit [DBG] from='client.? 192.168.123.102:0/699741649' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:24.282074+0000 mon.a (mon.0) 915 : audit [DBG] from='client.? 192.168.123.102:0/699741649' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:24.291194+0000 mon.a (mon.0) 916 : audit [DBG] from='client.? 192.168.123.102:0/3109639298' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:24.291194+0000 mon.a (mon.0) 916 : audit [DBG] from='client.? 192.168.123.102:0/3109639298' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:24.385821+0000 mon.a (mon.0) 917 : audit [DBG] from='client.? 192.168.123.102:0/1053561522' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:24.385821+0000 mon.a (mon.0) 917 : audit [DBG] from='client.? 192.168.123.102:0/1053561522' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:24.394158+0000 mon.a (mon.0) 918 : audit [DBG] from='client.? 192.168.123.102:0/2833210967' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:24 vm07 bash[19945]: audit 2026-03-21T06:59:24.394158+0000 mon.a (mon.0) 918 : audit [DBG] from='client.? 192.168.123.102:0/2833210967' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:23.792826+0000 mon.a (mon.0) 912 : audit [DBG] from='client.? 192.168.123.102:0/1991895665' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:23.792826+0000 mon.a (mon.0) 912 : audit [DBG] from='client.? 192.168.123.102:0/1991895665' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:23.803170+0000 mon.a (mon.0) 913 : audit [DBG] from='client.? 192.168.123.102:0/3010713827' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:23.803170+0000 mon.a (mon.0) 913 : audit [DBG] from='client.? 192.168.123.102:0/3010713827' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:24.254491+0000 mon.a (mon.0) 914 : audit [DBG] from='client.? 192.168.123.102:0/1281175297' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:24.254491+0000 mon.a (mon.0) 914 : audit [DBG] from='client.? 192.168.123.102:0/1281175297' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:24.282074+0000 mon.a (mon.0) 915 : audit [DBG] from='client.? 192.168.123.102:0/699741649' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:24.282074+0000 mon.a (mon.0) 915 : audit [DBG] from='client.? 192.168.123.102:0/699741649' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:24.291194+0000 mon.a (mon.0) 916 : audit [DBG] from='client.? 192.168.123.102:0/3109639298' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:24.291194+0000 mon.a (mon.0) 916 : audit [DBG] from='client.? 192.168.123.102:0/3109639298' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:24.385821+0000 mon.a (mon.0) 917 : audit [DBG] from='client.? 192.168.123.102:0/1053561522' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:24.385821+0000 mon.a (mon.0) 917 : audit [DBG] from='client.? 192.168.123.102:0/1053561522' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:24.394158+0000 mon.a (mon.0) 918 : audit [DBG] from='client.? 192.168.123.102:0/2833210967' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:24 vm04 bash[20194]: audit 2026-03-21T06:59:24.394158+0000 mon.a (mon.0) 918 : audit [DBG] from='client.? 192.168.123.102:0/2833210967' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.791 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:25] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.792 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:25] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.792 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:25] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:25.792 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:25] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:25.792 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:25] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.792 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:25] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.792 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:25] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.792 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:25] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:25.886 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:24.826840+0000 mon.a (mon.0) 919 : audit [DBG] from='client.? 192.168.123.102:0/638363844' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:25.886 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:24.826840+0000 mon.a (mon.0) 919 : audit [DBG] from='client.? 192.168.123.102:0/638363844' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:25.886 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:24.851129+0000 mon.a (mon.0) 920 : audit [DBG] from='client.? 192.168.123.102:0/922501011' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:24.851129+0000 mon.a (mon.0) 920 : audit [DBG] from='client.? 192.168.123.102:0/922501011' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:24.855773+0000 mon.b (mon.2) 25 : audit [DBG] from='client.? 192.168.123.102:0/3566121643' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:24.855773+0000 mon.b (mon.2) 25 : audit [DBG] from='client.? 192.168.123.102:0/3566121643' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:24.954797+0000 mon.c (mon.1) 41 : audit [DBG] from='client.? 192.168.123.102:0/3273106188' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:24.954797+0000 mon.c (mon.1) 41 : audit [DBG] from='client.? 192.168.123.102:0/3273106188' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:24.969388+0000 mon.a (mon.0) 921 : audit [DBG] from='client.? 192.168.123.102:0/3356881030' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:24.969388+0000 mon.a (mon.0) 921 : audit [DBG] from='client.? 192.168.123.102:0/3356881030' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:25.413136+0000 mon.a (mon.0) 922 : audit [DBG] from='client.? 192.168.123.102:0/3199464217' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:25.413136+0000 mon.a (mon.0) 922 : audit [DBG] from='client.? 192.168.123.102:0/3199464217' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:25.438673+0000 mon.a (mon.0) 923 : audit [DBG] from='client.? 192.168.123.102:0/2907950886' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:25.438673+0000 mon.a (mon.0) 923 : audit [DBG] from='client.? 192.168.123.102:0/2907950886' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:25.448616+0000 mon.a (mon.0) 924 : audit [DBG] from='client.? 192.168.123.102:0/2894582720' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:25.448616+0000 mon.a (mon.0) 924 : audit [DBG] from='client.? 192.168.123.102:0/2894582720' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:25.543976+0000 mon.a (mon.0) 925 : audit [DBG] from='client.? 192.168.123.102:0/2891429488' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:25.543976+0000 mon.a (mon.0) 925 : audit [DBG] from='client.? 192.168.123.102:0/2891429488' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:25.551785+0000 mon.a (mon.0) 926 : audit [DBG] from='client.? 192.168.123.102:0/2044615474' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: audit 2026-03-21T06:59:25.551785+0000 mon.a (mon.0) 926 : audit [DBG] from='client.? 192.168.123.102:0/2044615474' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: cluster 2026-03-21T06:59:25.640744+0000 mgr.x (mgr.14152) 502 : cluster [DBG] pgmap v397: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 28 op/s 2026-03-21T06:59:25.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:25 vm02 bash[17657]: cluster 2026-03-21T06:59:25.640744+0000 mgr.x (mgr.14152) 502 : cluster [DBG] pgmap v397: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 28 op/s 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:24.826840+0000 mon.a (mon.0) 919 : audit [DBG] from='client.? 192.168.123.102:0/638363844' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:24.826840+0000 mon.a (mon.0) 919 : audit [DBG] from='client.? 192.168.123.102:0/638363844' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:24.851129+0000 mon.a (mon.0) 920 : audit [DBG] from='client.? 192.168.123.102:0/922501011' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:24.851129+0000 mon.a (mon.0) 920 : audit [DBG] from='client.? 192.168.123.102:0/922501011' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:24.855773+0000 mon.b (mon.2) 25 : audit [DBG] from='client.? 192.168.123.102:0/3566121643' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:24.855773+0000 mon.b (mon.2) 25 : audit [DBG] from='client.? 192.168.123.102:0/3566121643' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:24.954797+0000 mon.c (mon.1) 41 : audit [DBG] from='client.? 192.168.123.102:0/3273106188' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:24.954797+0000 mon.c (mon.1) 41 : audit [DBG] from='client.? 192.168.123.102:0/3273106188' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:24.969388+0000 mon.a (mon.0) 921 : audit [DBG] from='client.? 192.168.123.102:0/3356881030' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:24.969388+0000 mon.a (mon.0) 921 : audit [DBG] from='client.? 192.168.123.102:0/3356881030' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:25.413136+0000 mon.a (mon.0) 922 : audit [DBG] from='client.? 192.168.123.102:0/3199464217' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:25.413136+0000 mon.a (mon.0) 922 : audit [DBG] from='client.? 192.168.123.102:0/3199464217' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:25.438673+0000 mon.a (mon.0) 923 : audit [DBG] from='client.? 192.168.123.102:0/2907950886' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:25.438673+0000 mon.a (mon.0) 923 : audit [DBG] from='client.? 192.168.123.102:0/2907950886' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:25.448616+0000 mon.a (mon.0) 924 : audit [DBG] from='client.? 192.168.123.102:0/2894582720' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:25.448616+0000 mon.a (mon.0) 924 : audit [DBG] from='client.? 192.168.123.102:0/2894582720' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:25.543976+0000 mon.a (mon.0) 925 : audit [DBG] from='client.? 192.168.123.102:0/2891429488' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:25.543976+0000 mon.a (mon.0) 925 : audit [DBG] from='client.? 192.168.123.102:0/2891429488' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:25.551785+0000 mon.a (mon.0) 926 : audit [DBG] from='client.? 192.168.123.102:0/2044615474' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: audit 2026-03-21T06:59:25.551785+0000 mon.a (mon.0) 926 : audit [DBG] from='client.? 192.168.123.102:0/2044615474' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: cluster 2026-03-21T06:59:25.640744+0000 mgr.x (mgr.14152) 502 : cluster [DBG] pgmap v397: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 28 op/s 2026-03-21T06:59:26.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:25 vm07 bash[19945]: cluster 2026-03-21T06:59:25.640744+0000 mgr.x (mgr.14152) 502 : cluster [DBG] pgmap v397: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 28 op/s 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:24.826840+0000 mon.a (mon.0) 919 : audit [DBG] from='client.? 192.168.123.102:0/638363844' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:24.826840+0000 mon.a (mon.0) 919 : audit [DBG] from='client.? 192.168.123.102:0/638363844' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:24.851129+0000 mon.a (mon.0) 920 : audit [DBG] from='client.? 192.168.123.102:0/922501011' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:24.851129+0000 mon.a (mon.0) 920 : audit [DBG] from='client.? 192.168.123.102:0/922501011' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:24.855773+0000 mon.b (mon.2) 25 : audit [DBG] from='client.? 192.168.123.102:0/3566121643' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:24.855773+0000 mon.b (mon.2) 25 : audit [DBG] from='client.? 192.168.123.102:0/3566121643' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:24.954797+0000 mon.c (mon.1) 41 : audit [DBG] from='client.? 192.168.123.102:0/3273106188' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:24.954797+0000 mon.c (mon.1) 41 : audit [DBG] from='client.? 192.168.123.102:0/3273106188' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:24.969388+0000 mon.a (mon.0) 921 : audit [DBG] from='client.? 192.168.123.102:0/3356881030' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:24.969388+0000 mon.a (mon.0) 921 : audit [DBG] from='client.? 192.168.123.102:0/3356881030' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:25.413136+0000 mon.a (mon.0) 922 : audit [DBG] from='client.? 192.168.123.102:0/3199464217' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:25.413136+0000 mon.a (mon.0) 922 : audit [DBG] from='client.? 192.168.123.102:0/3199464217' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:25.438673+0000 mon.a (mon.0) 923 : audit [DBG] from='client.? 192.168.123.102:0/2907950886' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:25.438673+0000 mon.a (mon.0) 923 : audit [DBG] from='client.? 192.168.123.102:0/2907950886' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:25.448616+0000 mon.a (mon.0) 924 : audit [DBG] from='client.? 192.168.123.102:0/2894582720' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:25.448616+0000 mon.a (mon.0) 924 : audit [DBG] from='client.? 192.168.123.102:0/2894582720' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:25.543976+0000 mon.a (mon.0) 925 : audit [DBG] from='client.? 192.168.123.102:0/2891429488' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:25.543976+0000 mon.a (mon.0) 925 : audit [DBG] from='client.? 192.168.123.102:0/2891429488' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:25.551785+0000 mon.a (mon.0) 926 : audit [DBG] from='client.? 192.168.123.102:0/2044615474' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: audit 2026-03-21T06:59:25.551785+0000 mon.a (mon.0) 926 : audit [DBG] from='client.? 192.168.123.102:0/2044615474' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: cluster 2026-03-21T06:59:25.640744+0000 mgr.x (mgr.14152) 502 : cluster [DBG] pgmap v397: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 28 op/s 2026-03-21T06:59:26.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:25 vm04 bash[20194]: cluster 2026-03-21T06:59:25.640744+0000 mgr.x (mgr.14152) 502 : cluster [DBG] pgmap v397: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 28 op/s 2026-03-21T06:59:26.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:26.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:26.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:26.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:26.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:26.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:26.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:26.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.002802+0000 mon.a (mon.0) 927 : audit [DBG] from='client.? 192.168.123.102:0/3908590815' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.002802+0000 mon.a (mon.0) 927 : audit [DBG] from='client.? 192.168.123.102:0/3908590815' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.029879+0000 mon.a (mon.0) 928 : audit [DBG] from='client.? 192.168.123.102:0/2510262350' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.029879+0000 mon.a (mon.0) 928 : audit [DBG] from='client.? 192.168.123.102:0/2510262350' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.040964+0000 mon.a (mon.0) 929 : audit [DBG] from='client.? 192.168.123.102:0/733692424' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.040964+0000 mon.a (mon.0) 929 : audit [DBG] from='client.? 192.168.123.102:0/733692424' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.140315+0000 mon.a (mon.0) 930 : audit [DBG] from='client.? 192.168.123.102:0/1919119727' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.140315+0000 mon.a (mon.0) 930 : audit [DBG] from='client.? 192.168.123.102:0/1919119727' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.149058+0000 mon.a (mon.0) 931 : audit [DBG] from='client.? 192.168.123.102:0/3918466769' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.149058+0000 mon.a (mon.0) 931 : audit [DBG] from='client.? 192.168.123.102:0/3918466769' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.603595+0000 mon.a (mon.0) 932 : audit [DBG] from='client.? 192.168.123.102:0/3979118478' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.603595+0000 mon.a (mon.0) 932 : audit [DBG] from='client.? 192.168.123.102:0/3979118478' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.628324+0000 mon.a (mon.0) 933 : audit [DBG] from='client.? 192.168.123.102:0/1026972457' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.628324+0000 mon.a (mon.0) 933 : audit [DBG] from='client.? 192.168.123.102:0/1026972457' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.636854+0000 mon.a (mon.0) 934 : audit [DBG] from='client.? 192.168.123.102:0/3222623796' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.636854+0000 mon.a (mon.0) 934 : audit [DBG] from='client.? 192.168.123.102:0/3222623796' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.728815+0000 mon.a (mon.0) 935 : audit [DBG] from='client.? 192.168.123.102:0/2061633777' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.728815+0000 mon.a (mon.0) 935 : audit [DBG] from='client.? 192.168.123.102:0/2061633777' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.736473+0000 mon.a (mon.0) 936 : audit [DBG] from='client.? 192.168.123.102:0/10034223' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[17657]: audit 2026-03-21T06:59:26.736473+0000 mon.a (mon.0) 936 : audit [DBG] from='client.? 192.168.123.102:0/10034223' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug gateway validation needed for vm02.local 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug An additional 1 tpg's are required 2026-03-21T06:59:27.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T06:59:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:59:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-21T06:59:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm02.local HTTP/1.1" 200 - 2026-03-21T06:59:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm02.local HTTP/1.1" 200 - 2026-03-21T06:59:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug _gateway update on localhost, successful 2026-03-21T06:59:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm02.local HTTP/1.1" 200 - 2026-03-21T06:59:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:26] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm02.local HTTP/1.1" 200 - 2026-03-21T06:59:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:26 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:26] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:27.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:26 vm07 bash[42142]: debug Change detected - internal 2 / xattr 3 refreshing 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.002802+0000 mon.a (mon.0) 927 : audit [DBG] from='client.? 192.168.123.102:0/3908590815' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.002802+0000 mon.a (mon.0) 927 : audit [DBG] from='client.? 192.168.123.102:0/3908590815' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.029879+0000 mon.a (mon.0) 928 : audit [DBG] from='client.? 192.168.123.102:0/2510262350' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.029879+0000 mon.a (mon.0) 928 : audit [DBG] from='client.? 192.168.123.102:0/2510262350' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.040964+0000 mon.a (mon.0) 929 : audit [DBG] from='client.? 192.168.123.102:0/733692424' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.040964+0000 mon.a (mon.0) 929 : audit [DBG] from='client.? 192.168.123.102:0/733692424' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.140315+0000 mon.a (mon.0) 930 : audit [DBG] from='client.? 192.168.123.102:0/1919119727' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.140315+0000 mon.a (mon.0) 930 : audit [DBG] from='client.? 192.168.123.102:0/1919119727' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.149058+0000 mon.a (mon.0) 931 : audit [DBG] from='client.? 192.168.123.102:0/3918466769' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.149058+0000 mon.a (mon.0) 931 : audit [DBG] from='client.? 192.168.123.102:0/3918466769' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.603595+0000 mon.a (mon.0) 932 : audit [DBG] from='client.? 192.168.123.102:0/3979118478' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.603595+0000 mon.a (mon.0) 932 : audit [DBG] from='client.? 192.168.123.102:0/3979118478' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.628324+0000 mon.a (mon.0) 933 : audit [DBG] from='client.? 192.168.123.102:0/1026972457' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.628324+0000 mon.a (mon.0) 933 : audit [DBG] from='client.? 192.168.123.102:0/1026972457' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.636854+0000 mon.a (mon.0) 934 : audit [DBG] from='client.? 192.168.123.102:0/3222623796' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.636854+0000 mon.a (mon.0) 934 : audit [DBG] from='client.? 192.168.123.102:0/3222623796' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.728815+0000 mon.a (mon.0) 935 : audit [DBG] from='client.? 192.168.123.102:0/2061633777' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.728815+0000 mon.a (mon.0) 935 : audit [DBG] from='client.? 192.168.123.102:0/2061633777' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.736473+0000 mon.a (mon.0) 936 : audit [DBG] from='client.? 192.168.123.102:0/10034223' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:26 vm07 bash[19945]: audit 2026-03-21T06:59:26.736473+0000 mon.a (mon.0) 936 : audit [DBG] from='client.? 192.168.123.102:0/10034223' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.002802+0000 mon.a (mon.0) 927 : audit [DBG] from='client.? 192.168.123.102:0/3908590815' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.002802+0000 mon.a (mon.0) 927 : audit [DBG] from='client.? 192.168.123.102:0/3908590815' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.029879+0000 mon.a (mon.0) 928 : audit [DBG] from='client.? 192.168.123.102:0/2510262350' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.029879+0000 mon.a (mon.0) 928 : audit [DBG] from='client.? 192.168.123.102:0/2510262350' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.040964+0000 mon.a (mon.0) 929 : audit [DBG] from='client.? 192.168.123.102:0/733692424' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.040964+0000 mon.a (mon.0) 929 : audit [DBG] from='client.? 192.168.123.102:0/733692424' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.140315+0000 mon.a (mon.0) 930 : audit [DBG] from='client.? 192.168.123.102:0/1919119727' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.140315+0000 mon.a (mon.0) 930 : audit [DBG] from='client.? 192.168.123.102:0/1919119727' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.149058+0000 mon.a (mon.0) 931 : audit [DBG] from='client.? 192.168.123.102:0/3918466769' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.149058+0000 mon.a (mon.0) 931 : audit [DBG] from='client.? 192.168.123.102:0/3918466769' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.603595+0000 mon.a (mon.0) 932 : audit [DBG] from='client.? 192.168.123.102:0/3979118478' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.603595+0000 mon.a (mon.0) 932 : audit [DBG] from='client.? 192.168.123.102:0/3979118478' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.628324+0000 mon.a (mon.0) 933 : audit [DBG] from='client.? 192.168.123.102:0/1026972457' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.628324+0000 mon.a (mon.0) 933 : audit [DBG] from='client.? 192.168.123.102:0/1026972457' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.636854+0000 mon.a (mon.0) 934 : audit [DBG] from='client.? 192.168.123.102:0/3222623796' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.636854+0000 mon.a (mon.0) 934 : audit [DBG] from='client.? 192.168.123.102:0/3222623796' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.728815+0000 mon.a (mon.0) 935 : audit [DBG] from='client.? 192.168.123.102:0/2061633777' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.728815+0000 mon.a (mon.0) 935 : audit [DBG] from='client.? 192.168.123.102:0/2061633777' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.736473+0000 mon.a (mon.0) 936 : audit [DBG] from='client.? 192.168.123.102:0/10034223' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:26 vm04 bash[20194]: audit 2026-03-21T06:59:26.736473+0000 mon.a (mon.0) 936 : audit [DBG] from='client.? 192.168.123.102:0/10034223' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:27] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:27] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:27] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:27.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:27] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:27.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:27] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:27] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:27] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:27] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:27.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:27.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:27.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:27.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:27] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:27.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:27 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:26.799382+0000 mgr.x (mgr.14152) 503 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:26.799382+0000 mgr.x (mgr.14152) 503 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.255807+0000 mon.a (mon.0) 937 : audit [DBG] from='client.? 192.168.123.102:0/3333579134' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.255807+0000 mon.a (mon.0) 937 : audit [DBG] from='client.? 192.168.123.102:0/3333579134' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.282384+0000 mon.a (mon.0) 938 : audit [DBG] from='client.? 192.168.123.102:0/3967419108' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.282384+0000 mon.a (mon.0) 938 : audit [DBG] from='client.? 192.168.123.102:0/3967419108' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.295876+0000 mon.a (mon.0) 939 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.295876+0000 mon.a (mon.0) 939 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.299693+0000 mon.a (mon.0) 940 : audit [DBG] from='client.? 192.168.123.102:0/3061904113' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.299693+0000 mon.a (mon.0) 940 : audit [DBG] from='client.? 192.168.123.102:0/3061904113' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.301763+0000 mon.a (mon.0) 941 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.301763+0000 mon.a (mon.0) 941 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.407352+0000 mon.a (mon.0) 942 : audit [DBG] from='client.? 192.168.123.102:0/751577587' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.407352+0000 mon.a (mon.0) 942 : audit [DBG] from='client.? 192.168.123.102:0/751577587' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.416547+0000 mon.a (mon.0) 943 : audit [DBG] from='client.? 192.168.123.102:0/4131151290' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.416547+0000 mon.a (mon.0) 943 : audit [DBG] from='client.? 192.168.123.102:0/4131151290' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.632694+0000 mgr.x (mgr.14152) 504 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.632694+0000 mgr.x (mgr.14152) 504 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: cluster 2026-03-21T06:59:27.641243+0000 mgr.x (mgr.14152) 505 : cluster [DBG] pgmap v398: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 27 op/s 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: cluster 2026-03-21T06:59:27.641243+0000 mgr.x (mgr.14152) 505 : cluster [DBG] pgmap v398: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 27 op/s 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.661517+0000 mon.a (mon.0) 944 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.661517+0000 mon.a (mon.0) 944 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.662460+0000 mon.a (mon.0) 945 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.662460+0000 mon.a (mon.0) 945 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.668830+0000 mon.a (mon.0) 946 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:27.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.668830+0000 mon.a (mon.0) 946 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:27.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.670021+0000 mon.a (mon.0) 947 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:59:27.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:27 vm07 bash[19945]: audit 2026-03-21T06:59:27.670021+0000 mon.a (mon.0) 947 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:27] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:27] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:28.057 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:26.799382+0000 mgr.x (mgr.14152) 503 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:28.057 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:26.799382+0000 mgr.x (mgr.14152) 503 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:28.057 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.255807+0000 mon.a (mon.0) 937 : audit [DBG] from='client.? 192.168.123.102:0/3333579134' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:28.057 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.255807+0000 mon.a (mon.0) 937 : audit [DBG] from='client.? 192.168.123.102:0/3333579134' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:28.057 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.282384+0000 mon.a (mon.0) 938 : audit [DBG] from='client.? 192.168.123.102:0/3967419108' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:28.057 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.282384+0000 mon.a (mon.0) 938 : audit [DBG] from='client.? 192.168.123.102:0/3967419108' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:28.057 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.295876+0000 mon.a (mon.0) 939 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.057 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.295876+0000 mon.a (mon.0) 939 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.299693+0000 mon.a (mon.0) 940 : audit [DBG] from='client.? 192.168.123.102:0/3061904113' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.299693+0000 mon.a (mon.0) 940 : audit [DBG] from='client.? 192.168.123.102:0/3061904113' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.301763+0000 mon.a (mon.0) 941 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.301763+0000 mon.a (mon.0) 941 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.407352+0000 mon.a (mon.0) 942 : audit [DBG] from='client.? 192.168.123.102:0/751577587' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.407352+0000 mon.a (mon.0) 942 : audit [DBG] from='client.? 192.168.123.102:0/751577587' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.416547+0000 mon.a (mon.0) 943 : audit [DBG] from='client.? 192.168.123.102:0/4131151290' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.416547+0000 mon.a (mon.0) 943 : audit [DBG] from='client.? 192.168.123.102:0/4131151290' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.632694+0000 mgr.x (mgr.14152) 504 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.632694+0000 mgr.x (mgr.14152) 504 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: cluster 2026-03-21T06:59:27.641243+0000 mgr.x (mgr.14152) 505 : cluster [DBG] pgmap v398: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 27 op/s 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: cluster 2026-03-21T06:59:27.641243+0000 mgr.x (mgr.14152) 505 : cluster [DBG] pgmap v398: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 27 op/s 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.661517+0000 mon.a (mon.0) 944 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.661517+0000 mon.a (mon.0) 944 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.662460+0000 mon.a (mon.0) 945 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.662460+0000 mon.a (mon.0) 945 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.668830+0000 mon.a (mon.0) 946 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.668830+0000 mon.a (mon.0) 946 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.670021+0000 mon.a (mon.0) 947 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:59:28.058 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:27 vm02 bash[17657]: audit 2026-03-21T06:59:27.670021+0000 mon.a (mon.0) 947 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:26.799382+0000 mgr.x (mgr.14152) 503 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:26.799382+0000 mgr.x (mgr.14152) 503 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.255807+0000 mon.a (mon.0) 937 : audit [DBG] from='client.? 192.168.123.102:0/3333579134' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.255807+0000 mon.a (mon.0) 937 : audit [DBG] from='client.? 192.168.123.102:0/3333579134' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.282384+0000 mon.a (mon.0) 938 : audit [DBG] from='client.? 192.168.123.102:0/3967419108' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.282384+0000 mon.a (mon.0) 938 : audit [DBG] from='client.? 192.168.123.102:0/3967419108' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.295876+0000 mon.a (mon.0) 939 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.295876+0000 mon.a (mon.0) 939 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.299693+0000 mon.a (mon.0) 940 : audit [DBG] from='client.? 192.168.123.102:0/3061904113' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.299693+0000 mon.a (mon.0) 940 : audit [DBG] from='client.? 192.168.123.102:0/3061904113' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.301763+0000 mon.a (mon.0) 941 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.301763+0000 mon.a (mon.0) 941 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.407352+0000 mon.a (mon.0) 942 : audit [DBG] from='client.? 192.168.123.102:0/751577587' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:28.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.407352+0000 mon.a (mon.0) 942 : audit [DBG] from='client.? 192.168.123.102:0/751577587' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.416547+0000 mon.a (mon.0) 943 : audit [DBG] from='client.? 192.168.123.102:0/4131151290' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.416547+0000 mon.a (mon.0) 943 : audit [DBG] from='client.? 192.168.123.102:0/4131151290' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.632694+0000 mgr.x (mgr.14152) 504 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.632694+0000 mgr.x (mgr.14152) 504 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: cluster 2026-03-21T06:59:27.641243+0000 mgr.x (mgr.14152) 505 : cluster [DBG] pgmap v398: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 27 op/s 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: cluster 2026-03-21T06:59:27.641243+0000 mgr.x (mgr.14152) 505 : cluster [DBG] pgmap v398: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 22 KiB/s rd, 597 B/s wr, 27 op/s 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.661517+0000 mon.a (mon.0) 944 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.661517+0000 mon.a (mon.0) 944 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.662460+0000 mon.a (mon.0) 945 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.662460+0000 mon.a (mon.0) 945 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.668830+0000 mon.a (mon.0) 946 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.668830+0000 mon.a (mon.0) 946 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.670021+0000 mon.a (mon.0) 947 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:59:28.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:27 vm04 bash[20194]: audit 2026-03-21T06:59:27.670021+0000 mon.a (mon.0) 947 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug gateway validation needed for vm07.local 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug _gateway update on vm07.local, successful 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug An additional 1 tpg's are required 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug _gateway update on vm02.local, successful 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T06:59:28.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/sysinfo/hostname HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:28.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:28.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:28.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:28] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:27.902508+0000 mon.b (mon.2) 26 : audit [DBG] from='client.? 192.168.123.102:0/2294966163' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:27.902508+0000 mon.b (mon.2) 26 : audit [DBG] from='client.? 192.168.123.102:0/2294966163' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:27.930547+0000 mon.c (mon.1) 42 : audit [DBG] from='client.? 192.168.123.102:0/958783521' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:27.930547+0000 mon.c (mon.1) 42 : audit [DBG] from='client.? 192.168.123.102:0/958783521' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:27.945712+0000 mon.a (mon.0) 948 : audit [DBG] from='client.? 192.168.123.102:0/963117732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:27.945712+0000 mon.a (mon.0) 948 : audit [DBG] from='client.? 192.168.123.102:0/963117732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.046293+0000 mon.a (mon.0) 949 : audit [DBG] from='client.? 192.168.123.102:0/941695149' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.046293+0000 mon.a (mon.0) 949 : audit [DBG] from='client.? 192.168.123.102:0/941695149' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.054528+0000 mon.a (mon.0) 950 : audit [DBG] from='client.? 192.168.123.102:0/3282903783' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.054528+0000 mon.a (mon.0) 950 : audit [DBG] from='client.? 192.168.123.102:0/3282903783' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.624098+0000 mon.a (mon.0) 951 : audit [DBG] from='client.? 192.168.123.102:0/2290764894' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.624098+0000 mon.a (mon.0) 951 : audit [DBG] from='client.? 192.168.123.102:0/2290764894' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.644461+0000 mon.a (mon.0) 952 : audit [DBG] from='client.? 192.168.123.102:0/359422249' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.644461+0000 mon.a (mon.0) 952 : audit [DBG] from='client.? 192.168.123.102:0/359422249' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.652392+0000 mon.a (mon.0) 953 : audit [DBG] from='client.? 192.168.123.102:0/1105765677' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.652392+0000 mon.a (mon.0) 953 : audit [DBG] from='client.? 192.168.123.102:0/1105765677' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.747573+0000 mon.a (mon.0) 954 : audit [DBG] from='client.? 192.168.123.102:0/72057106' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.747573+0000 mon.a (mon.0) 954 : audit [DBG] from='client.? 192.168.123.102:0/72057106' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.754312+0000 mon.a (mon.0) 955 : audit [DBG] from='client.? 192.168.123.102:0/1013822341' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:28 vm02 bash[17657]: audit 2026-03-21T06:59:28.754312+0000 mon.a (mon.0) 955 : audit [DBG] from='client.? 192.168.123.102:0/1013822341' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:28 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:28] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:27.902508+0000 mon.b (mon.2) 26 : audit [DBG] from='client.? 192.168.123.102:0/2294966163' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:27.902508+0000 mon.b (mon.2) 26 : audit [DBG] from='client.? 192.168.123.102:0/2294966163' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:27.930547+0000 mon.c (mon.1) 42 : audit [DBG] from='client.? 192.168.123.102:0/958783521' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:27.930547+0000 mon.c (mon.1) 42 : audit [DBG] from='client.? 192.168.123.102:0/958783521' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:27.945712+0000 mon.a (mon.0) 948 : audit [DBG] from='client.? 192.168.123.102:0/963117732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:27.945712+0000 mon.a (mon.0) 948 : audit [DBG] from='client.? 192.168.123.102:0/963117732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.046293+0000 mon.a (mon.0) 949 : audit [DBG] from='client.? 192.168.123.102:0/941695149' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.046293+0000 mon.a (mon.0) 949 : audit [DBG] from='client.? 192.168.123.102:0/941695149' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.054528+0000 mon.a (mon.0) 950 : audit [DBG] from='client.? 192.168.123.102:0/3282903783' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.054528+0000 mon.a (mon.0) 950 : audit [DBG] from='client.? 192.168.123.102:0/3282903783' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.624098+0000 mon.a (mon.0) 951 : audit [DBG] from='client.? 192.168.123.102:0/2290764894' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.624098+0000 mon.a (mon.0) 951 : audit [DBG] from='client.? 192.168.123.102:0/2290764894' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.644461+0000 mon.a (mon.0) 952 : audit [DBG] from='client.? 192.168.123.102:0/359422249' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.644461+0000 mon.a (mon.0) 952 : audit [DBG] from='client.? 192.168.123.102:0/359422249' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.652392+0000 mon.a (mon.0) 953 : audit [DBG] from='client.? 192.168.123.102:0/1105765677' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.652392+0000 mon.a (mon.0) 953 : audit [DBG] from='client.? 192.168.123.102:0/1105765677' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.747573+0000 mon.a (mon.0) 954 : audit [DBG] from='client.? 192.168.123.102:0/72057106' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.747573+0000 mon.a (mon.0) 954 : audit [DBG] from='client.? 192.168.123.102:0/72057106' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.754312+0000 mon.a (mon.0) 955 : audit [DBG] from='client.? 192.168.123.102:0/1013822341' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:28 vm07 bash[19945]: audit 2026-03-21T06:59:28.754312+0000 mon.a (mon.0) 955 : audit [DBG] from='client.? 192.168.123.102:0/1013822341' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:27.902508+0000 mon.b (mon.2) 26 : audit [DBG] from='client.? 192.168.123.102:0/2294966163' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:27.902508+0000 mon.b (mon.2) 26 : audit [DBG] from='client.? 192.168.123.102:0/2294966163' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:27.930547+0000 mon.c (mon.1) 42 : audit [DBG] from='client.? 192.168.123.102:0/958783521' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:27.930547+0000 mon.c (mon.1) 42 : audit [DBG] from='client.? 192.168.123.102:0/958783521' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:27.945712+0000 mon.a (mon.0) 948 : audit [DBG] from='client.? 192.168.123.102:0/963117732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:27.945712+0000 mon.a (mon.0) 948 : audit [DBG] from='client.? 192.168.123.102:0/963117732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.046293+0000 mon.a (mon.0) 949 : audit [DBG] from='client.? 192.168.123.102:0/941695149' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.046293+0000 mon.a (mon.0) 949 : audit [DBG] from='client.? 192.168.123.102:0/941695149' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.054528+0000 mon.a (mon.0) 950 : audit [DBG] from='client.? 192.168.123.102:0/3282903783' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.054528+0000 mon.a (mon.0) 950 : audit [DBG] from='client.? 192.168.123.102:0/3282903783' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.624098+0000 mon.a (mon.0) 951 : audit [DBG] from='client.? 192.168.123.102:0/2290764894' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.624098+0000 mon.a (mon.0) 951 : audit [DBG] from='client.? 192.168.123.102:0/2290764894' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.644461+0000 mon.a (mon.0) 952 : audit [DBG] from='client.? 192.168.123.102:0/359422249' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.644461+0000 mon.a (mon.0) 952 : audit [DBG] from='client.? 192.168.123.102:0/359422249' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.652392+0000 mon.a (mon.0) 953 : audit [DBG] from='client.? 192.168.123.102:0/1105765677' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.652392+0000 mon.a (mon.0) 953 : audit [DBG] from='client.? 192.168.123.102:0/1105765677' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.747573+0000 mon.a (mon.0) 954 : audit [DBG] from='client.? 192.168.123.102:0/72057106' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.747573+0000 mon.a (mon.0) 954 : audit [DBG] from='client.? 192.168.123.102:0/72057106' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.754312+0000 mon.a (mon.0) 955 : audit [DBG] from='client.? 192.168.123.102:0/1013822341' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:28 vm04 bash[20194]: audit 2026-03-21T06:59:28.754312+0000 mon.a (mon.0) 955 : audit [DBG] from='client.? 192.168.123.102:0/1013822341' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block0' to LIO backstore user:rbd 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug (LUN.add_dev_to_lio) Successfully added datapool/block0 to LIO 2026-03-21T06:59:29.442 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T06:59:29.529 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.529 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.529 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.529 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.529 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.529 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug Setup group ano2 for datapool.block0 on tpg 2 (state 1, owner False, failover type 1) 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug Setup group ao for datapool.block0 on tpg 1 (state 0, owner True, failover type 1) 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug _targetlun update on vm02.local, successful 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug _targetlun update on vm07.local, successful 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:29] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block0' to LIO backstore user:rbd 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: /usr/lib/python3.9/site-packages/rtslib_fb/root.py:180: UserWarning: Cannot set dbroot to /var/target. Target devices have already been registered. 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: warn("Cannot set dbroot to {}. Target devices have already been registered." 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: debug (LUN.add_dev_to_lio) Successfully added datapool/block0 to LIO 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: debug (LUN.allocate) added block0 to LIO using wwn '36001405da17b74481464e9fa968746d3' defined by vm02.local 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: debug Setup group ano2 for datapool.block0 on tpg 2 (state 1, owner False, failover type 1) 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: debug Setup group ao for datapool.block0 on tpg 1 (state 0, owner True, failover type 1) 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:29.819 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:29 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:29] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:29.886 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.158649+0000 mon.a (mon.0) 956 : audit [DBG] from='client.? 192.168.123.102:0/2619237397' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.158649+0000 mon.a (mon.0) 956 : audit [DBG] from='client.? 192.168.123.102:0/2619237397' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.179550+0000 mon.a (mon.0) 957 : audit [DBG] from='client.? 192.168.123.102:0/2205401057' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.179550+0000 mon.a (mon.0) 957 : audit [DBG] from='client.? 192.168.123.102:0/2205401057' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.187108+0000 mon.a (mon.0) 958 : audit [DBG] from='client.? 192.168.123.102:0/3420064652' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.187108+0000 mon.a (mon.0) 958 : audit [DBG] from='client.? 192.168.123.102:0/3420064652' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.284254+0000 mon.a (mon.0) 959 : audit [DBG] from='client.? 192.168.123.102:0/308758637' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.284254+0000 mon.a (mon.0) 959 : audit [DBG] from='client.? 192.168.123.102:0/308758637' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.291339+0000 mon.a (mon.0) 960 : audit [DBG] from='client.? 192.168.123.102:0/2855989259' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.291339+0000 mon.a (mon.0) 960 : audit [DBG] from='client.? 192.168.123.102:0/2855989259' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.430102+0000 mon.a (mon.0) 961 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.430102+0000 mon.a (mon.0) 961 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.430921+0000 mon.a (mon.0) 962 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.430921+0000 mon.a (mon.0) 962 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.431234+0000 mon.a (mon.0) 963 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.431234+0000 mon.a (mon.0) 963 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.438715+0000 mon.a (mon.0) 964 : audit [INF] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2640075721"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.438715+0000 mon.a (mon.0) 964 : audit [INF] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2640075721"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.545419+0000 mon.c (mon.1) 43 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.545419+0000 mon.c (mon.1) 43 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.546192+0000 mon.c (mon.1) 44 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.546192+0000 mon.c (mon.1) 44 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.546489+0000 mon.c (mon.1) 45 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: audit 2026-03-21T06:59:29.546489+0000 mon.c (mon.1) 45 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: cluster 2026-03-21T06:59:29.641675+0000 mgr.x (mgr.14152) 506 : cluster [DBG] pgmap v399: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 597 B/s wr, 45 op/s 2026-03-21T06:59:29.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:29 vm02 bash[17657]: cluster 2026-03-21T06:59:29.641675+0000 mgr.x (mgr.14152) 506 : cluster [DBG] pgmap v399: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 597 B/s wr, 45 op/s 2026-03-21T06:59:30.077 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.077 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.158649+0000 mon.a (mon.0) 956 : audit [DBG] from='client.? 192.168.123.102:0/2619237397' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.158649+0000 mon.a (mon.0) 956 : audit [DBG] from='client.? 192.168.123.102:0/2619237397' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.179550+0000 mon.a (mon.0) 957 : audit [DBG] from='client.? 192.168.123.102:0/2205401057' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.179550+0000 mon.a (mon.0) 957 : audit [DBG] from='client.? 192.168.123.102:0/2205401057' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.187108+0000 mon.a (mon.0) 958 : audit [DBG] from='client.? 192.168.123.102:0/3420064652' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.187108+0000 mon.a (mon.0) 958 : audit [DBG] from='client.? 192.168.123.102:0/3420064652' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.284254+0000 mon.a (mon.0) 959 : audit [DBG] from='client.? 192.168.123.102:0/308758637' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.284254+0000 mon.a (mon.0) 959 : audit [DBG] from='client.? 192.168.123.102:0/308758637' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.291339+0000 mon.a (mon.0) 960 : audit [DBG] from='client.? 192.168.123.102:0/2855989259' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.291339+0000 mon.a (mon.0) 960 : audit [DBG] from='client.? 192.168.123.102:0/2855989259' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.430102+0000 mon.a (mon.0) 961 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.430102+0000 mon.a (mon.0) 961 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.430921+0000 mon.a (mon.0) 962 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.430921+0000 mon.a (mon.0) 962 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.431234+0000 mon.a (mon.0) 963 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.431234+0000 mon.a (mon.0) 963 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.438715+0000 mon.a (mon.0) 964 : audit [INF] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2640075721"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.438715+0000 mon.a (mon.0) 964 : audit [INF] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2640075721"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.545419+0000 mon.c (mon.1) 43 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.545419+0000 mon.c (mon.1) 43 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.546192+0000 mon.c (mon.1) 44 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.546192+0000 mon.c (mon.1) 44 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.546489+0000 mon.c (mon.1) 45 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: audit 2026-03-21T06:59:29.546489+0000 mon.c (mon.1) 45 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: cluster 2026-03-21T06:59:29.641675+0000 mgr.x (mgr.14152) 506 : cluster [DBG] pgmap v399: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 597 B/s wr, 45 op/s 2026-03-21T06:59:30.098 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:29 vm07 bash[19945]: cluster 2026-03-21T06:59:29.641675+0000 mgr.x (mgr.14152) 506 : cluster [DBG] pgmap v399: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 597 B/s wr, 45 op/s 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.158649+0000 mon.a (mon.0) 956 : audit [DBG] from='client.? 192.168.123.102:0/2619237397' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.158649+0000 mon.a (mon.0) 956 : audit [DBG] from='client.? 192.168.123.102:0/2619237397' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.179550+0000 mon.a (mon.0) 957 : audit [DBG] from='client.? 192.168.123.102:0/2205401057' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.179550+0000 mon.a (mon.0) 957 : audit [DBG] from='client.? 192.168.123.102:0/2205401057' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.187108+0000 mon.a (mon.0) 958 : audit [DBG] from='client.? 192.168.123.102:0/3420064652' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.187108+0000 mon.a (mon.0) 958 : audit [DBG] from='client.? 192.168.123.102:0/3420064652' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.284254+0000 mon.a (mon.0) 959 : audit [DBG] from='client.? 192.168.123.102:0/308758637' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.284254+0000 mon.a (mon.0) 959 : audit [DBG] from='client.? 192.168.123.102:0/308758637' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.291339+0000 mon.a (mon.0) 960 : audit [DBG] from='client.? 192.168.123.102:0/2855989259' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.291339+0000 mon.a (mon.0) 960 : audit [DBG] from='client.? 192.168.123.102:0/2855989259' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.430102+0000 mon.a (mon.0) 961 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.430102+0000 mon.a (mon.0) 961 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.430921+0000 mon.a (mon.0) 962 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.430921+0000 mon.a (mon.0) 962 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.431234+0000 mon.a (mon.0) 963 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.431234+0000 mon.a (mon.0) 963 : audit [DBG] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.438715+0000 mon.a (mon.0) 964 : audit [INF] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2640075721"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.438715+0000 mon.a (mon.0) 964 : audit [INF] from='client.? 192.168.123.102:0/2538308856' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2640075721"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.545419+0000 mon.c (mon.1) 43 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.545419+0000 mon.c (mon.1) 43 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.546192+0000 mon.c (mon.1) 44 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.546192+0000 mon.c (mon.1) 44 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.546489+0000 mon.c (mon.1) 45 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: audit 2026-03-21T06:59:29.546489+0000 mon.c (mon.1) 45 : audit [DBG] from='client.? 192.168.123.107:0/3792060432' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: cluster 2026-03-21T06:59:29.641675+0000 mgr.x (mgr.14152) 506 : cluster [DBG] pgmap v399: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 597 B/s wr, 45 op/s 2026-03-21T06:59:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:29 vm04 bash[20194]: cluster 2026-03-21T06:59:29.641675+0000 mgr.x (mgr.14152) 506 : cluster [DBG] pgmap v399: 4 pgs: 4 active+clean; 450 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 36 KiB/s rd, 597 B/s wr, 45 op/s 2026-03-21T06:59:30.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:30.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:30.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug _client update on localhost, successful 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug _client update on vm07.local, successful 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.829 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:30] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:30.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:30.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-21T06:59:30.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-21T06:59:30.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:30.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:30 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:30] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:29.986065+0000 mon.a (mon.0) 965 : audit [DBG] from='client.? 192.168.123.102:0/3211939350' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:29.986065+0000 mon.a (mon.0) 965 : audit [DBG] from='client.? 192.168.123.102:0/3211939350' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.008251+0000 mon.a (mon.0) 966 : audit [DBG] from='client.? 192.168.123.102:0/1670895308' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.008251+0000 mon.a (mon.0) 966 : audit [DBG] from='client.? 192.168.123.102:0/1670895308' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.017311+0000 mon.a (mon.0) 967 : audit [DBG] from='client.? 192.168.123.102:0/1173362757' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.017311+0000 mon.a (mon.0) 967 : audit [DBG] from='client.? 192.168.123.102:0/1173362757' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.114372+0000 mon.a (mon.0) 968 : audit [DBG] from='client.? 192.168.123.102:0/357404069' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.114372+0000 mon.a (mon.0) 968 : audit [DBG] from='client.? 192.168.123.102:0/357404069' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.121879+0000 mon.a (mon.0) 969 : audit [DBG] from='client.? 192.168.123.102:0/924288594' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.121879+0000 mon.a (mon.0) 969 : audit [DBG] from='client.? 192.168.123.102:0/924288594' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.521089+0000 mon.a (mon.0) 970 : audit [DBG] from='client.? 192.168.123.102:0/528907924' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.521089+0000 mon.a (mon.0) 970 : audit [DBG] from='client.? 192.168.123.102:0/528907924' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.541512+0000 mon.a (mon.0) 971 : audit [DBG] from='client.? 192.168.123.102:0/1789399841' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.541512+0000 mon.a (mon.0) 971 : audit [DBG] from='client.? 192.168.123.102:0/1789399841' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.549257+0000 mon.a (mon.0) 972 : audit [DBG] from='client.? 192.168.123.102:0/1968354415' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.549257+0000 mon.a (mon.0) 972 : audit [DBG] from='client.? 192.168.123.102:0/1968354415' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.669083+0000 mon.a (mon.0) 973 : audit [DBG] from='client.? 192.168.123.102:0/397636787' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.669083+0000 mon.a (mon.0) 973 : audit [DBG] from='client.? 192.168.123.102:0/397636787' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.680695+0000 mon.a (mon.0) 974 : audit [DBG] from='client.? 192.168.123.102:0/4138383805' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:30.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:30 vm07 bash[19945]: audit 2026-03-21T06:59:30.680695+0000 mon.a (mon.0) 974 : audit [DBG] from='client.? 192.168.123.102:0/4138383805' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:29.986065+0000 mon.a (mon.0) 965 : audit [DBG] from='client.? 192.168.123.102:0/3211939350' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:29.986065+0000 mon.a (mon.0) 965 : audit [DBG] from='client.? 192.168.123.102:0/3211939350' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.008251+0000 mon.a (mon.0) 966 : audit [DBG] from='client.? 192.168.123.102:0/1670895308' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.008251+0000 mon.a (mon.0) 966 : audit [DBG] from='client.? 192.168.123.102:0/1670895308' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.017311+0000 mon.a (mon.0) 967 : audit [DBG] from='client.? 192.168.123.102:0/1173362757' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.017311+0000 mon.a (mon.0) 967 : audit [DBG] from='client.? 192.168.123.102:0/1173362757' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.114372+0000 mon.a (mon.0) 968 : audit [DBG] from='client.? 192.168.123.102:0/357404069' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.114372+0000 mon.a (mon.0) 968 : audit [DBG] from='client.? 192.168.123.102:0/357404069' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.121879+0000 mon.a (mon.0) 969 : audit [DBG] from='client.? 192.168.123.102:0/924288594' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.121879+0000 mon.a (mon.0) 969 : audit [DBG] from='client.? 192.168.123.102:0/924288594' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.521089+0000 mon.a (mon.0) 970 : audit [DBG] from='client.? 192.168.123.102:0/528907924' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.521089+0000 mon.a (mon.0) 970 : audit [DBG] from='client.? 192.168.123.102:0/528907924' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.541512+0000 mon.a (mon.0) 971 : audit [DBG] from='client.? 192.168.123.102:0/1789399841' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.541512+0000 mon.a (mon.0) 971 : audit [DBG] from='client.? 192.168.123.102:0/1789399841' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.549257+0000 mon.a (mon.0) 972 : audit [DBG] from='client.? 192.168.123.102:0/1968354415' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.549257+0000 mon.a (mon.0) 972 : audit [DBG] from='client.? 192.168.123.102:0/1968354415' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.669083+0000 mon.a (mon.0) 973 : audit [DBG] from='client.? 192.168.123.102:0/397636787' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.669083+0000 mon.a (mon.0) 973 : audit [DBG] from='client.? 192.168.123.102:0/397636787' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.680695+0000 mon.a (mon.0) 974 : audit [DBG] from='client.? 192.168.123.102:0/4138383805' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:30 vm02 bash[17657]: audit 2026-03-21T06:59:30.680695+0000 mon.a (mon.0) 974 : audit [DBG] from='client.? 192.168.123.102:0/4138383805' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:31.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:29.986065+0000 mon.a (mon.0) 965 : audit [DBG] from='client.? 192.168.123.102:0/3211939350' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:31.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:29.986065+0000 mon.a (mon.0) 965 : audit [DBG] from='client.? 192.168.123.102:0/3211939350' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:31.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.008251+0000 mon.a (mon.0) 966 : audit [DBG] from='client.? 192.168.123.102:0/1670895308' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.008251+0000 mon.a (mon.0) 966 : audit [DBG] from='client.? 192.168.123.102:0/1670895308' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.017311+0000 mon.a (mon.0) 967 : audit [DBG] from='client.? 192.168.123.102:0/1173362757' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.017311+0000 mon.a (mon.0) 967 : audit [DBG] from='client.? 192.168.123.102:0/1173362757' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.114372+0000 mon.a (mon.0) 968 : audit [DBG] from='client.? 192.168.123.102:0/357404069' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.114372+0000 mon.a (mon.0) 968 : audit [DBG] from='client.? 192.168.123.102:0/357404069' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.121879+0000 mon.a (mon.0) 969 : audit [DBG] from='client.? 192.168.123.102:0/924288594' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.121879+0000 mon.a (mon.0) 969 : audit [DBG] from='client.? 192.168.123.102:0/924288594' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.521089+0000 mon.a (mon.0) 970 : audit [DBG] from='client.? 192.168.123.102:0/528907924' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.521089+0000 mon.a (mon.0) 970 : audit [DBG] from='client.? 192.168.123.102:0/528907924' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.541512+0000 mon.a (mon.0) 971 : audit [DBG] from='client.? 192.168.123.102:0/1789399841' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.541512+0000 mon.a (mon.0) 971 : audit [DBG] from='client.? 192.168.123.102:0/1789399841' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.549257+0000 mon.a (mon.0) 972 : audit [DBG] from='client.? 192.168.123.102:0/1968354415' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.549257+0000 mon.a (mon.0) 972 : audit [DBG] from='client.? 192.168.123.102:0/1968354415' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.669083+0000 mon.a (mon.0) 973 : audit [DBG] from='client.? 192.168.123.102:0/397636787' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.669083+0000 mon.a (mon.0) 973 : audit [DBG] from='client.? 192.168.123.102:0/397636787' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.680695+0000 mon.a (mon.0) 974 : audit [DBG] from='client.? 192.168.123.102:0/4138383805' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:30 vm04 bash[20194]: audit 2026-03-21T06:59:30.680695+0000 mon.a (mon.0) 974 : audit [DBG] from='client.? 192.168.123.102:0/4138383805' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:31.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:31.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:31.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:31.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:31.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:31.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:31.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:31.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:31.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:31.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:31.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:31.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:31.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:31 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:31.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:31 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:31.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:31 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:31.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:31 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:31] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.223575+0000 mon.a (mon.0) 975 : audit [DBG] from='client.? 192.168.123.102:0/50160528' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.223575+0000 mon.a (mon.0) 975 : audit [DBG] from='client.? 192.168.123.102:0/50160528' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.249706+0000 mon.a (mon.0) 976 : audit [DBG] from='client.? 192.168.123.102:0/4095797799' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.249706+0000 mon.a (mon.0) 976 : audit [DBG] from='client.? 192.168.123.102:0/4095797799' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.260713+0000 mon.a (mon.0) 977 : audit [DBG] from='client.? 192.168.123.102:0/1730685216' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.260713+0000 mon.a (mon.0) 977 : audit [DBG] from='client.? 192.168.123.102:0/1730685216' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.376223+0000 mon.a (mon.0) 978 : audit [DBG] from='client.? 192.168.123.102:0/983777732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.376223+0000 mon.a (mon.0) 978 : audit [DBG] from='client.? 192.168.123.102:0/983777732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.384249+0000 mon.a (mon.0) 979 : audit [DBG] from='client.? 192.168.123.102:0/4276565360' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.384249+0000 mon.a (mon.0) 979 : audit [DBG] from='client.? 192.168.123.102:0/4276565360' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: cluster 2026-03-21T06:59:31.642071+0000 mgr.x (mgr.14152) 507 : cluster [DBG] pgmap v400: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 1.3 KiB/s wr, 60 op/s 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: cluster 2026-03-21T06:59:31.642071+0000 mgr.x (mgr.14152) 507 : cluster [DBG] pgmap v400: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 1.3 KiB/s wr, 60 op/s 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.830520+0000 mon.a (mon.0) 980 : audit [DBG] from='client.? 192.168.123.102:0/2301778247' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:31 vm02 bash[17657]: audit 2026-03-21T06:59:31.830520+0000 mon.a (mon.0) 980 : audit [DBG] from='client.? 192.168.123.102:0/2301778247' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:31 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:31 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:31 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:31 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:31] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.223575+0000 mon.a (mon.0) 975 : audit [DBG] from='client.? 192.168.123.102:0/50160528' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.223575+0000 mon.a (mon.0) 975 : audit [DBG] from='client.? 192.168.123.102:0/50160528' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.249706+0000 mon.a (mon.0) 976 : audit [DBG] from='client.? 192.168.123.102:0/4095797799' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.249706+0000 mon.a (mon.0) 976 : audit [DBG] from='client.? 192.168.123.102:0/4095797799' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.260713+0000 mon.a (mon.0) 977 : audit [DBG] from='client.? 192.168.123.102:0/1730685216' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.260713+0000 mon.a (mon.0) 977 : audit [DBG] from='client.? 192.168.123.102:0/1730685216' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.376223+0000 mon.a (mon.0) 978 : audit [DBG] from='client.? 192.168.123.102:0/983777732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.376223+0000 mon.a (mon.0) 978 : audit [DBG] from='client.? 192.168.123.102:0/983777732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.384249+0000 mon.a (mon.0) 979 : audit [DBG] from='client.? 192.168.123.102:0/4276565360' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.384249+0000 mon.a (mon.0) 979 : audit [DBG] from='client.? 192.168.123.102:0/4276565360' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: cluster 2026-03-21T06:59:31.642071+0000 mgr.x (mgr.14152) 507 : cluster [DBG] pgmap v400: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 1.3 KiB/s wr, 60 op/s 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: cluster 2026-03-21T06:59:31.642071+0000 mgr.x (mgr.14152) 507 : cluster [DBG] pgmap v400: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 1.3 KiB/s wr, 60 op/s 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.830520+0000 mon.a (mon.0) 980 : audit [DBG] from='client.? 192.168.123.102:0/2301778247' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:31 vm07 bash[19945]: audit 2026-03-21T06:59:31.830520+0000 mon.a (mon.0) 980 : audit [DBG] from='client.? 192.168.123.102:0/2301778247' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.223575+0000 mon.a (mon.0) 975 : audit [DBG] from='client.? 192.168.123.102:0/50160528' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.223575+0000 mon.a (mon.0) 975 : audit [DBG] from='client.? 192.168.123.102:0/50160528' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.249706+0000 mon.a (mon.0) 976 : audit [DBG] from='client.? 192.168.123.102:0/4095797799' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.249706+0000 mon.a (mon.0) 976 : audit [DBG] from='client.? 192.168.123.102:0/4095797799' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.260713+0000 mon.a (mon.0) 977 : audit [DBG] from='client.? 192.168.123.102:0/1730685216' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.260713+0000 mon.a (mon.0) 977 : audit [DBG] from='client.? 192.168.123.102:0/1730685216' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.376223+0000 mon.a (mon.0) 978 : audit [DBG] from='client.? 192.168.123.102:0/983777732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.376223+0000 mon.a (mon.0) 978 : audit [DBG] from='client.? 192.168.123.102:0/983777732' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.384249+0000 mon.a (mon.0) 979 : audit [DBG] from='client.? 192.168.123.102:0/4276565360' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:32.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.384249+0000 mon.a (mon.0) 979 : audit [DBG] from='client.? 192.168.123.102:0/4276565360' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:32.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: cluster 2026-03-21T06:59:31.642071+0000 mgr.x (mgr.14152) 507 : cluster [DBG] pgmap v400: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 1.3 KiB/s wr, 60 op/s 2026-03-21T06:59:32.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: cluster 2026-03-21T06:59:31.642071+0000 mgr.x (mgr.14152) 507 : cluster [DBG] pgmap v400: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 53 KiB/s rd, 1.3 KiB/s wr, 60 op/s 2026-03-21T06:59:32.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.830520+0000 mon.a (mon.0) 980 : audit [DBG] from='client.? 192.168.123.102:0/2301778247' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:31 vm04 bash[20194]: audit 2026-03-21T06:59:31.830520+0000 mon.a (mon.0) 980 : audit [DBG] from='client.? 192.168.123.102:0/2301778247' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug (Client.add_lun) added image 'datapool/block0' to iqn.1994-05.com.redhat:client 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug _clientlun update on localhost, successful 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug _clientlun update on vm07.local, successful 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:32.708 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:32] "GET /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:32.846 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:32 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.846 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:32 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.846 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:32 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.846 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:32 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.846 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:32 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.846 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:32 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:32.846 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:32 vm07 bash[42142]: debug (Client.add_lun) added image 'datapool/block0' to iqn.1994-05.com.redhat:client 2026-03-21T06:59:32.846 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:32 vm07 bash[42142]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-21T06:59:32.846 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:32 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:32.846 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:32 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:32] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:33.115 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:31.859483+0000 mon.a (mon.0) 981 : audit [DBG] from='client.? 192.168.123.102:0/8003388' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:31.859483+0000 mon.a (mon.0) 981 : audit [DBG] from='client.? 192.168.123.102:0/8003388' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:31.869146+0000 mon.a (mon.0) 982 : audit [DBG] from='client.? 192.168.123.102:0/759597883' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:31.869146+0000 mon.a (mon.0) 982 : audit [DBG] from='client.? 192.168.123.102:0/759597883' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:31.982505+0000 mon.a (mon.0) 983 : audit [DBG] from='client.? 192.168.123.102:0/1632235901' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:31.982505+0000 mon.a (mon.0) 983 : audit [DBG] from='client.? 192.168.123.102:0/1632235901' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:31.990112+0000 mon.a (mon.0) 984 : audit [DBG] from='client.? 192.168.123.102:0/3678701392' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:31.990112+0000 mon.a (mon.0) 984 : audit [DBG] from='client.? 192.168.123.102:0/3678701392' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:32.418943+0000 mon.a (mon.0) 985 : audit [DBG] from='client.? 192.168.123.102:0/2026775037' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:32.418943+0000 mon.a (mon.0) 985 : audit [DBG] from='client.? 192.168.123.102:0/2026775037' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:32.445630+0000 mon.a (mon.0) 986 : audit [DBG] from='client.? 192.168.123.102:0/3202120309' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:32.445630+0000 mon.a (mon.0) 986 : audit [DBG] from='client.? 192.168.123.102:0/3202120309' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:32.455355+0000 mon.a (mon.0) 987 : audit [DBG] from='client.? 192.168.123.102:0/498474174' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:32.455355+0000 mon.a (mon.0) 987 : audit [DBG] from='client.? 192.168.123.102:0/498474174' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:32.571807+0000 mon.a (mon.0) 988 : audit [DBG] from='client.? 192.168.123.102:0/4273100716' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:32.571807+0000 mon.a (mon.0) 988 : audit [DBG] from='client.? 192.168.123.102:0/4273100716' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:32.582285+0000 mon.a (mon.0) 989 : audit [DBG] from='client.? 192.168.123.102:0/2142122532' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:32 vm02 bash[17657]: audit 2026-03-21T06:59:32.582285+0000 mon.a (mon.0) 989 : audit [DBG] from='client.? 192.168.123.102:0/2142122532' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:31.859483+0000 mon.a (mon.0) 981 : audit [DBG] from='client.? 192.168.123.102:0/8003388' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:31.859483+0000 mon.a (mon.0) 981 : audit [DBG] from='client.? 192.168.123.102:0/8003388' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:31.869146+0000 mon.a (mon.0) 982 : audit [DBG] from='client.? 192.168.123.102:0/759597883' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:31.869146+0000 mon.a (mon.0) 982 : audit [DBG] from='client.? 192.168.123.102:0/759597883' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:31.982505+0000 mon.a (mon.0) 983 : audit [DBG] from='client.? 192.168.123.102:0/1632235901' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:31.982505+0000 mon.a (mon.0) 983 : audit [DBG] from='client.? 192.168.123.102:0/1632235901' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:31.990112+0000 mon.a (mon.0) 984 : audit [DBG] from='client.? 192.168.123.102:0/3678701392' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:31.990112+0000 mon.a (mon.0) 984 : audit [DBG] from='client.? 192.168.123.102:0/3678701392' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:32.418943+0000 mon.a (mon.0) 985 : audit [DBG] from='client.? 192.168.123.102:0/2026775037' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:32.418943+0000 mon.a (mon.0) 985 : audit [DBG] from='client.? 192.168.123.102:0/2026775037' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:32.445630+0000 mon.a (mon.0) 986 : audit [DBG] from='client.? 192.168.123.102:0/3202120309' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:32.445630+0000 mon.a (mon.0) 986 : audit [DBG] from='client.? 192.168.123.102:0/3202120309' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:32.455355+0000 mon.a (mon.0) 987 : audit [DBG] from='client.? 192.168.123.102:0/498474174' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:32.455355+0000 mon.a (mon.0) 987 : audit [DBG] from='client.? 192.168.123.102:0/498474174' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:32.571807+0000 mon.a (mon.0) 988 : audit [DBG] from='client.? 192.168.123.102:0/4273100716' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:32.571807+0000 mon.a (mon.0) 988 : audit [DBG] from='client.? 192.168.123.102:0/4273100716' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:32.582285+0000 mon.a (mon.0) 989 : audit [DBG] from='client.? 192.168.123.102:0/2142122532' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:32 vm07 bash[19945]: audit 2026-03-21T06:59:32.582285+0000 mon.a (mon.0) 989 : audit [DBG] from='client.? 192.168.123.102:0/2142122532' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:31.859483+0000 mon.a (mon.0) 981 : audit [DBG] from='client.? 192.168.123.102:0/8003388' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:31.859483+0000 mon.a (mon.0) 981 : audit [DBG] from='client.? 192.168.123.102:0/8003388' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:31.869146+0000 mon.a (mon.0) 982 : audit [DBG] from='client.? 192.168.123.102:0/759597883' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:31.869146+0000 mon.a (mon.0) 982 : audit [DBG] from='client.? 192.168.123.102:0/759597883' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:31.982505+0000 mon.a (mon.0) 983 : audit [DBG] from='client.? 192.168.123.102:0/1632235901' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:31.982505+0000 mon.a (mon.0) 983 : audit [DBG] from='client.? 192.168.123.102:0/1632235901' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:31.990112+0000 mon.a (mon.0) 984 : audit [DBG] from='client.? 192.168.123.102:0/3678701392' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:31.990112+0000 mon.a (mon.0) 984 : audit [DBG] from='client.? 192.168.123.102:0/3678701392' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:32.418943+0000 mon.a (mon.0) 985 : audit [DBG] from='client.? 192.168.123.102:0/2026775037' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:32.418943+0000 mon.a (mon.0) 985 : audit [DBG] from='client.? 192.168.123.102:0/2026775037' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:32.445630+0000 mon.a (mon.0) 986 : audit [DBG] from='client.? 192.168.123.102:0/3202120309' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:32.445630+0000 mon.a (mon.0) 986 : audit [DBG] from='client.? 192.168.123.102:0/3202120309' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:32.455355+0000 mon.a (mon.0) 987 : audit [DBG] from='client.? 192.168.123.102:0/498474174' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:32.455355+0000 mon.a (mon.0) 987 : audit [DBG] from='client.? 192.168.123.102:0/498474174' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:32.571807+0000 mon.a (mon.0) 988 : audit [DBG] from='client.? 192.168.123.102:0/4273100716' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:32.571807+0000 mon.a (mon.0) 988 : audit [DBG] from='client.? 192.168.123.102:0/4273100716' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:32.582285+0000 mon.a (mon.0) 989 : audit [DBG] from='client.? 192.168.123.102:0/2142122532' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:32 vm04 bash[20194]: audit 2026-03-21T06:59:32.582285+0000 mon.a (mon.0) 989 : audit [DBG] from='client.? 192.168.123.102:0/2142122532' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:33.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:33.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:33.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:33.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:33.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:33.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:33.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:33.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:33.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:33.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:33.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:33.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:33.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:33 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:33.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:33 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:33.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:33 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:33.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:33 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:33.775 INFO:tasks.cram.client.0.vm02.stdout:/home/ubuntu/cephtest/archive/cram.client.0/gwcli_create.t: passed 2026-03-21T06:59:33.775 INFO:tasks.cram.client.0.vm02.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-21T06:59:33.780 INFO:tasks.cram:Running tests for client.1... 2026-03-21T06:59:33.780 DEBUG:teuthology.orchestra.run.vm04:> CEPH_REF=master CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.1/*.t 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:33] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:34.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.084676+0000 mon.a (mon.0) 990 : audit [DBG] from='client.? 192.168.123.102:0/1682584611' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.084676+0000 mon.a (mon.0) 990 : audit [DBG] from='client.? 192.168.123.102:0/1682584611' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.106570+0000 mon.a (mon.0) 991 : audit [DBG] from='client.? 192.168.123.102:0/611796848' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.106570+0000 mon.a (mon.0) 991 : audit [DBG] from='client.? 192.168.123.102:0/611796848' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.115024+0000 mon.a (mon.0) 992 : audit [DBG] from='client.? 192.168.123.102:0/3761924445' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.115024+0000 mon.a (mon.0) 992 : audit [DBG] from='client.? 192.168.123.102:0/3761924445' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.214875+0000 mon.a (mon.0) 993 : audit [DBG] from='client.? 192.168.123.102:0/886849393' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.214875+0000 mon.a (mon.0) 993 : audit [DBG] from='client.? 192.168.123.102:0/886849393' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.222117+0000 mon.a (mon.0) 994 : audit [DBG] from='client.? 192.168.123.102:0/371698730' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.222117+0000 mon.a (mon.0) 994 : audit [DBG] from='client.? 192.168.123.102:0/371698730' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.618360+0000 mon.a (mon.0) 995 : audit [DBG] from='client.? 192.168.123.102:0/3514079077' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.618360+0000 mon.a (mon.0) 995 : audit [DBG] from='client.? 192.168.123.102:0/3514079077' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.632624+0000 mon.b (mon.2) 27 : audit [DBG] from='client.? 192.168.123.102:0/4040581014' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.632624+0000 mon.b (mon.2) 27 : audit [DBG] from='client.? 192.168.123.102:0/4040581014' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: cluster 2026-03-21T06:59:33.642358+0000 mgr.x (mgr.14152) 508 : cluster [DBG] pgmap v401: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 51 KiB/s rd, 1.1 KiB/s wr, 56 op/s 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: cluster 2026-03-21T06:59:33.642358+0000 mgr.x (mgr.14152) 508 : cluster [DBG] pgmap v401: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 51 KiB/s rd, 1.1 KiB/s wr, 56 op/s 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.647119+0000 mon.a (mon.0) 996 : audit [DBG] from='client.? 192.168.123.102:0/2667890872' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.647119+0000 mon.a (mon.0) 996 : audit [DBG] from='client.? 192.168.123.102:0/2667890872' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.736601+0000 mon.b (mon.2) 28 : audit [DBG] from='client.? 192.168.123.102:0/2182286137' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.736601+0000 mon.b (mon.2) 28 : audit [DBG] from='client.? 192.168.123.102:0/2182286137' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.744367+0000 mon.b (mon.2) 29 : audit [DBG] from='client.? 192.168.123.102:0/3453591171' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:34.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:33 vm02 bash[17657]: audit 2026-03-21T06:59:33.744367+0000 mon.b (mon.2) 29 : audit [DBG] from='client.? 192.168.123.102:0/3453591171' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:33 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:34.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:33 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:34.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:33 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:34.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:33 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 06:59:33] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.084676+0000 mon.a (mon.0) 990 : audit [DBG] from='client.? 192.168.123.102:0/1682584611' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.084676+0000 mon.a (mon.0) 990 : audit [DBG] from='client.? 192.168.123.102:0/1682584611' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.106570+0000 mon.a (mon.0) 991 : audit [DBG] from='client.? 192.168.123.102:0/611796848' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.106570+0000 mon.a (mon.0) 991 : audit [DBG] from='client.? 192.168.123.102:0/611796848' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.115024+0000 mon.a (mon.0) 992 : audit [DBG] from='client.? 192.168.123.102:0/3761924445' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.115024+0000 mon.a (mon.0) 992 : audit [DBG] from='client.? 192.168.123.102:0/3761924445' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.214875+0000 mon.a (mon.0) 993 : audit [DBG] from='client.? 192.168.123.102:0/886849393' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.214875+0000 mon.a (mon.0) 993 : audit [DBG] from='client.? 192.168.123.102:0/886849393' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.222117+0000 mon.a (mon.0) 994 : audit [DBG] from='client.? 192.168.123.102:0/371698730' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.222117+0000 mon.a (mon.0) 994 : audit [DBG] from='client.? 192.168.123.102:0/371698730' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.618360+0000 mon.a (mon.0) 995 : audit [DBG] from='client.? 192.168.123.102:0/3514079077' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.618360+0000 mon.a (mon.0) 995 : audit [DBG] from='client.? 192.168.123.102:0/3514079077' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.632624+0000 mon.b (mon.2) 27 : audit [DBG] from='client.? 192.168.123.102:0/4040581014' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.632624+0000 mon.b (mon.2) 27 : audit [DBG] from='client.? 192.168.123.102:0/4040581014' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: cluster 2026-03-21T06:59:33.642358+0000 mgr.x (mgr.14152) 508 : cluster [DBG] pgmap v401: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 51 KiB/s rd, 1.1 KiB/s wr, 56 op/s 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: cluster 2026-03-21T06:59:33.642358+0000 mgr.x (mgr.14152) 508 : cluster [DBG] pgmap v401: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 51 KiB/s rd, 1.1 KiB/s wr, 56 op/s 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.647119+0000 mon.a (mon.0) 996 : audit [DBG] from='client.? 192.168.123.102:0/2667890872' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.647119+0000 mon.a (mon.0) 996 : audit [DBG] from='client.? 192.168.123.102:0/2667890872' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.736601+0000 mon.b (mon.2) 28 : audit [DBG] from='client.? 192.168.123.102:0/2182286137' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.736601+0000 mon.b (mon.2) 28 : audit [DBG] from='client.? 192.168.123.102:0/2182286137' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.744367+0000 mon.b (mon.2) 29 : audit [DBG] from='client.? 192.168.123.102:0/3453591171' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:34.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:33 vm07 bash[19945]: audit 2026-03-21T06:59:33.744367+0000 mon.b (mon.2) 29 : audit [DBG] from='client.? 192.168.123.102:0/3453591171' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.084676+0000 mon.a (mon.0) 990 : audit [DBG] from='client.? 192.168.123.102:0/1682584611' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.084676+0000 mon.a (mon.0) 990 : audit [DBG] from='client.? 192.168.123.102:0/1682584611' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.106570+0000 mon.a (mon.0) 991 : audit [DBG] from='client.? 192.168.123.102:0/611796848' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.106570+0000 mon.a (mon.0) 991 : audit [DBG] from='client.? 192.168.123.102:0/611796848' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.115024+0000 mon.a (mon.0) 992 : audit [DBG] from='client.? 192.168.123.102:0/3761924445' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.115024+0000 mon.a (mon.0) 992 : audit [DBG] from='client.? 192.168.123.102:0/3761924445' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.214875+0000 mon.a (mon.0) 993 : audit [DBG] from='client.? 192.168.123.102:0/886849393' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.214875+0000 mon.a (mon.0) 993 : audit [DBG] from='client.? 192.168.123.102:0/886849393' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.222117+0000 mon.a (mon.0) 994 : audit [DBG] from='client.? 192.168.123.102:0/371698730' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.222117+0000 mon.a (mon.0) 994 : audit [DBG] from='client.? 192.168.123.102:0/371698730' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.618360+0000 mon.a (mon.0) 995 : audit [DBG] from='client.? 192.168.123.102:0/3514079077' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.618360+0000 mon.a (mon.0) 995 : audit [DBG] from='client.? 192.168.123.102:0/3514079077' entity='client.iscsi.iscsi.a' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.632624+0000 mon.b (mon.2) 27 : audit [DBG] from='client.? 192.168.123.102:0/4040581014' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.632624+0000 mon.b (mon.2) 27 : audit [DBG] from='client.? 192.168.123.102:0/4040581014' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: cluster 2026-03-21T06:59:33.642358+0000 mgr.x (mgr.14152) 508 : cluster [DBG] pgmap v401: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 51 KiB/s rd, 1.1 KiB/s wr, 56 op/s 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: cluster 2026-03-21T06:59:33.642358+0000 mgr.x (mgr.14152) 508 : cluster [DBG] pgmap v401: 4 pgs: 4 active+clean; 451 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 51 KiB/s rd, 1.1 KiB/s wr, 56 op/s 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.647119+0000 mon.a (mon.0) 996 : audit [DBG] from='client.? 192.168.123.102:0/2667890872' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.647119+0000 mon.a (mon.0) 996 : audit [DBG] from='client.? 192.168.123.102:0/2667890872' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.736601+0000 mon.b (mon.2) 28 : audit [DBG] from='client.? 192.168.123.102:0/2182286137' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.736601+0000 mon.b (mon.2) 28 : audit [DBG] from='client.? 192.168.123.102:0/2182286137' entity='client.iscsi.iscsi.a' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.744367+0000 mon.b (mon.2) 29 : audit [DBG] from='client.? 192.168.123.102:0/3453591171' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:34.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:33 vm04 bash[20194]: audit 2026-03-21T06:59:33.744367+0000 mon.b (mon.2) 29 : audit [DBG] from='client.? 192.168.123.102:0/3453591171' entity='client.iscsi.iscsi.a' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:37.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:36 vm04 bash[20194]: cluster 2026-03-21T06:59:35.642688+0000 mgr.x (mgr.14152) 509 : cluster [DBG] pgmap v402: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 96 KiB/s rd, 2.1 KiB/s wr, 102 op/s 2026-03-21T06:59:37.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:36 vm04 bash[20194]: cluster 2026-03-21T06:59:35.642688+0000 mgr.x (mgr.14152) 509 : cluster [DBG] pgmap v402: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 96 KiB/s rd, 2.1 KiB/s wr, 102 op/s 2026-03-21T06:59:37.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:36 vm02 bash[17657]: cluster 2026-03-21T06:59:35.642688+0000 mgr.x (mgr.14152) 509 : cluster [DBG] pgmap v402: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 96 KiB/s rd, 2.1 KiB/s wr, 102 op/s 2026-03-21T06:59:37.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:36 vm02 bash[17657]: cluster 2026-03-21T06:59:35.642688+0000 mgr.x (mgr.14152) 509 : cluster [DBG] pgmap v402: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 96 KiB/s rd, 2.1 KiB/s wr, 102 op/s 2026-03-21T06:59:37.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:36 vm07 bash[19945]: cluster 2026-03-21T06:59:35.642688+0000 mgr.x (mgr.14152) 509 : cluster [DBG] pgmap v402: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 96 KiB/s rd, 2.1 KiB/s wr, 102 op/s 2026-03-21T06:59:37.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:36 vm07 bash[19945]: cluster 2026-03-21T06:59:35.642688+0000 mgr.x (mgr.14152) 509 : cluster [DBG] pgmap v402: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 96 KiB/s rd, 2.1 KiB/s wr, 102 op/s 2026-03-21T06:59:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:38 vm04 bash[20194]: audit 2026-03-21T06:59:36.809863+0000 mgr.x (mgr.14152) 510 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:38 vm04 bash[20194]: audit 2026-03-21T06:59:36.809863+0000 mgr.x (mgr.14152) 510 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:38 vm04 bash[20194]: cluster 2026-03-21T06:59:37.642991+0000 mgr.x (mgr.14152) 511 : cluster [DBG] pgmap v403: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 76 KiB/s rd, 1.7 KiB/s wr, 78 op/s 2026-03-21T06:59:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:38 vm04 bash[20194]: cluster 2026-03-21T06:59:37.642991+0000 mgr.x (mgr.14152) 511 : cluster [DBG] pgmap v403: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 76 KiB/s rd, 1.7 KiB/s wr, 78 op/s 2026-03-21T06:59:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:38 vm04 bash[20194]: audit 2026-03-21T06:59:37.643539+0000 mgr.x (mgr.14152) 512 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:38 vm04 bash[20194]: audit 2026-03-21T06:59:37.643539+0000 mgr.x (mgr.14152) 512 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:39.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:38 vm02 bash[17657]: audit 2026-03-21T06:59:36.809863+0000 mgr.x (mgr.14152) 510 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:39.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:38 vm02 bash[17657]: audit 2026-03-21T06:59:36.809863+0000 mgr.x (mgr.14152) 510 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:39.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:38 vm02 bash[17657]: cluster 2026-03-21T06:59:37.642991+0000 mgr.x (mgr.14152) 511 : cluster [DBG] pgmap v403: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 76 KiB/s rd, 1.7 KiB/s wr, 78 op/s 2026-03-21T06:59:39.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:38 vm02 bash[17657]: cluster 2026-03-21T06:59:37.642991+0000 mgr.x (mgr.14152) 511 : cluster [DBG] pgmap v403: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 76 KiB/s rd, 1.7 KiB/s wr, 78 op/s 2026-03-21T06:59:39.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:38 vm02 bash[17657]: audit 2026-03-21T06:59:37.643539+0000 mgr.x (mgr.14152) 512 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:39.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:38 vm02 bash[17657]: audit 2026-03-21T06:59:37.643539+0000 mgr.x (mgr.14152) 512 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:38 vm07 bash[19945]: audit 2026-03-21T06:59:36.809863+0000 mgr.x (mgr.14152) 510 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:38 vm07 bash[19945]: audit 2026-03-21T06:59:36.809863+0000 mgr.x (mgr.14152) 510 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:38 vm07 bash[19945]: cluster 2026-03-21T06:59:37.642991+0000 mgr.x (mgr.14152) 511 : cluster [DBG] pgmap v403: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 76 KiB/s rd, 1.7 KiB/s wr, 78 op/s 2026-03-21T06:59:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:38 vm07 bash[19945]: cluster 2026-03-21T06:59:37.642991+0000 mgr.x (mgr.14152) 511 : cluster [DBG] pgmap v403: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 76 KiB/s rd, 1.7 KiB/s wr, 78 op/s 2026-03-21T06:59:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:38 vm07 bash[19945]: audit 2026-03-21T06:59:37.643539+0000 mgr.x (mgr.14152) 512 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:38 vm07 bash[19945]: audit 2026-03-21T06:59:37.643539+0000 mgr.x (mgr.14152) 512 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:41.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:40 vm04 bash[20194]: cluster 2026-03-21T06:59:39.643349+0000 mgr.x (mgr.14152) 513 : cluster [DBG] pgmap v404: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 78 KiB/s rd, 1.7 KiB/s wr, 80 op/s 2026-03-21T06:59:41.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:40 vm04 bash[20194]: cluster 2026-03-21T06:59:39.643349+0000 mgr.x (mgr.14152) 513 : cluster [DBG] pgmap v404: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 78 KiB/s rd, 1.7 KiB/s wr, 80 op/s 2026-03-21T06:59:41.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:40 vm02 bash[17657]: cluster 2026-03-21T06:59:39.643349+0000 mgr.x (mgr.14152) 513 : cluster [DBG] pgmap v404: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 78 KiB/s rd, 1.7 KiB/s wr, 80 op/s 2026-03-21T06:59:41.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:40 vm02 bash[17657]: cluster 2026-03-21T06:59:39.643349+0000 mgr.x (mgr.14152) 513 : cluster [DBG] pgmap v404: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 78 KiB/s rd, 1.7 KiB/s wr, 80 op/s 2026-03-21T06:59:41.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:40 vm07 bash[19945]: cluster 2026-03-21T06:59:39.643349+0000 mgr.x (mgr.14152) 513 : cluster [DBG] pgmap v404: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 78 KiB/s rd, 1.7 KiB/s wr, 80 op/s 2026-03-21T06:59:41.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:40 vm07 bash[19945]: cluster 2026-03-21T06:59:39.643349+0000 mgr.x (mgr.14152) 513 : cluster [DBG] pgmap v404: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 78 KiB/s rd, 1.7 KiB/s wr, 80 op/s 2026-03-21T06:59:43.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:42 vm04 bash[20194]: cluster 2026-03-21T06:59:41.643714+0000 mgr.x (mgr.14152) 514 : cluster [DBG] pgmap v405: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 64 KiB/s rd, 1.7 KiB/s wr, 63 op/s 2026-03-21T06:59:43.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:42 vm04 bash[20194]: cluster 2026-03-21T06:59:41.643714+0000 mgr.x (mgr.14152) 514 : cluster [DBG] pgmap v405: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 64 KiB/s rd, 1.7 KiB/s wr, 63 op/s 2026-03-21T06:59:43.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:42 vm02 bash[17657]: cluster 2026-03-21T06:59:41.643714+0000 mgr.x (mgr.14152) 514 : cluster [DBG] pgmap v405: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 64 KiB/s rd, 1.7 KiB/s wr, 63 op/s 2026-03-21T06:59:43.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:42 vm02 bash[17657]: cluster 2026-03-21T06:59:41.643714+0000 mgr.x (mgr.14152) 514 : cluster [DBG] pgmap v405: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 64 KiB/s rd, 1.7 KiB/s wr, 63 op/s 2026-03-21T06:59:43.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:42 vm07 bash[19945]: cluster 2026-03-21T06:59:41.643714+0000 mgr.x (mgr.14152) 514 : cluster [DBG] pgmap v405: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 64 KiB/s rd, 1.7 KiB/s wr, 63 op/s 2026-03-21T06:59:43.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:42 vm07 bash[19945]: cluster 2026-03-21T06:59:41.643714+0000 mgr.x (mgr.14152) 514 : cluster [DBG] pgmap v405: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 64 KiB/s rd, 1.7 KiB/s wr, 63 op/s 2026-03-21T06:59:44.131 INFO:tasks.cram.client.1.vm04.stdout:/home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t: passed 2026-03-21T06:59:44.131 INFO:tasks.cram.client.1.vm04.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-21T06:59:44.135 INFO:tasks.cram:Running tests for client.2... 2026-03-21T06:59:44.135 DEBUG:teuthology.orchestra.run.vm07:> CEPH_REF=master CEPH_ID="2" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.2/*.t 2026-03-21T06:59:44.984 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:44 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:44] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:44.984 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:44 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:44] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:44.984 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:44 vm07 bash[19945]: cluster 2026-03-21T06:59:43.644022+0000 mgr.x (mgr.14152) 515 : cluster [DBG] pgmap v406: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 47 KiB/s rd, 1023 B/s wr, 49 op/s 2026-03-21T06:59:44.984 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:44 vm07 bash[19945]: cluster 2026-03-21T06:59:43.644022+0000 mgr.x (mgr.14152) 515 : cluster [DBG] pgmap v406: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 47 KiB/s rd, 1023 B/s wr, 49 op/s 2026-03-21T06:59:45.005 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:44 vm02 bash[17657]: cluster 2026-03-21T06:59:43.644022+0000 mgr.x (mgr.14152) 515 : cluster [DBG] pgmap v406: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 47 KiB/s rd, 1023 B/s wr, 49 op/s 2026-03-21T06:59:45.005 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:44 vm02 bash[17657]: cluster 2026-03-21T06:59:43.644022+0000 mgr.x (mgr.14152) 515 : cluster [DBG] pgmap v406: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 47 KiB/s rd, 1023 B/s wr, 49 op/s 2026-03-21T06:59:45.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:44 vm04 bash[20194]: cluster 2026-03-21T06:59:43.644022+0000 mgr.x (mgr.14152) 515 : cluster [DBG] pgmap v406: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 47 KiB/s rd, 1023 B/s wr, 49 op/s 2026-03-21T06:59:45.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:44 vm04 bash[20194]: cluster 2026-03-21T06:59:43.644022+0000 mgr.x (mgr.14152) 515 : cluster [DBG] pgmap v406: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 47 KiB/s rd, 1023 B/s wr, 49 op/s 2026-03-21T06:59:45.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:44 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:44] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:45.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:44 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:44] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:45.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:44 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:44] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:44 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:44] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:45.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug _client update on vm02.local, successful 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug _client update on localhost, successful 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:45 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:45] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:44.882493+0000 mon.c (mon.1) 46 : audit [DBG] from='client.? 192.168.123.107:0/1856718767' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:44.882493+0000 mon.c (mon.1) 46 : audit [DBG] from='client.? 192.168.123.107:0/1856718767' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:44.903791+0000 mon.c (mon.1) 47 : audit [DBG] from='client.? 192.168.123.107:0/570228412' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:44.903791+0000 mon.c (mon.1) 47 : audit [DBG] from='client.? 192.168.123.107:0/570228412' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:44.911915+0000 mon.c (mon.1) 48 : audit [DBG] from='client.? 192.168.123.107:0/2452525968' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:44.911915+0000 mon.c (mon.1) 48 : audit [DBG] from='client.? 192.168.123.107:0/2452525968' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.019793+0000 mon.c (mon.1) 49 : audit [DBG] from='client.? 192.168.123.107:0/3069139021' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.019793+0000 mon.c (mon.1) 49 : audit [DBG] from='client.? 192.168.123.107:0/3069139021' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.026861+0000 mon.c (mon.1) 50 : audit [DBG] from='client.? 192.168.123.107:0/4165205942' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.026861+0000 mon.c (mon.1) 50 : audit [DBG] from='client.? 192.168.123.107:0/4165205942' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.473840+0000 mon.b (mon.2) 30 : audit [DBG] from='client.? 192.168.123.107:0/1347450667' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:45.770 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.473840+0000 mon.b (mon.2) 30 : audit [DBG] from='client.? 192.168.123.107:0/1347450667' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:45.771 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.501113+0000 mon.c (mon.1) 51 : audit [DBG] from='client.? 192.168.123.107:0/2220525325' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:45.771 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.501113+0000 mon.c (mon.1) 51 : audit [DBG] from='client.? 192.168.123.107:0/2220525325' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:45.771 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.510884+0000 mon.c (mon.1) 52 : audit [DBG] from='client.? 192.168.123.107:0/4276385491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.771 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.510884+0000 mon.c (mon.1) 52 : audit [DBG] from='client.? 192.168.123.107:0/4276385491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.771 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.628746+0000 mon.c (mon.1) 53 : audit [DBG] from='client.? 192.168.123.107:0/886655627' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.771 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.628746+0000 mon.c (mon.1) 53 : audit [DBG] from='client.? 192.168.123.107:0/886655627' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.771 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.636588+0000 mon.b (mon.2) 31 : audit [DBG] from='client.? 192.168.123.107:0/969698203' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:45.771 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:45 vm07 bash[19945]: audit 2026-03-21T06:59:45.636588+0000 mon.b (mon.2) 31 : audit [DBG] from='client.? 192.168.123.107:0/969698203' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:45.886 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.886 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.886 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.886 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.886 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.886 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:45.886 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-21T06:59:45.886 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:45.886 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:45] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:44.882493+0000 mon.c (mon.1) 46 : audit [DBG] from='client.? 192.168.123.107:0/1856718767' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:44.882493+0000 mon.c (mon.1) 46 : audit [DBG] from='client.? 192.168.123.107:0/1856718767' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:44.903791+0000 mon.c (mon.1) 47 : audit [DBG] from='client.? 192.168.123.107:0/570228412' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:44.903791+0000 mon.c (mon.1) 47 : audit [DBG] from='client.? 192.168.123.107:0/570228412' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:44.911915+0000 mon.c (mon.1) 48 : audit [DBG] from='client.? 192.168.123.107:0/2452525968' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:44.911915+0000 mon.c (mon.1) 48 : audit [DBG] from='client.? 192.168.123.107:0/2452525968' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.019793+0000 mon.c (mon.1) 49 : audit [DBG] from='client.? 192.168.123.107:0/3069139021' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.019793+0000 mon.c (mon.1) 49 : audit [DBG] from='client.? 192.168.123.107:0/3069139021' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.026861+0000 mon.c (mon.1) 50 : audit [DBG] from='client.? 192.168.123.107:0/4165205942' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.026861+0000 mon.c (mon.1) 50 : audit [DBG] from='client.? 192.168.123.107:0/4165205942' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.473840+0000 mon.b (mon.2) 30 : audit [DBG] from='client.? 192.168.123.107:0/1347450667' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.473840+0000 mon.b (mon.2) 30 : audit [DBG] from='client.? 192.168.123.107:0/1347450667' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.501113+0000 mon.c (mon.1) 51 : audit [DBG] from='client.? 192.168.123.107:0/2220525325' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.501113+0000 mon.c (mon.1) 51 : audit [DBG] from='client.? 192.168.123.107:0/2220525325' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.510884+0000 mon.c (mon.1) 52 : audit [DBG] from='client.? 192.168.123.107:0/4276385491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.510884+0000 mon.c (mon.1) 52 : audit [DBG] from='client.? 192.168.123.107:0/4276385491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.628746+0000 mon.c (mon.1) 53 : audit [DBG] from='client.? 192.168.123.107:0/886655627' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.628746+0000 mon.c (mon.1) 53 : audit [DBG] from='client.? 192.168.123.107:0/886655627' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.636588+0000 mon.b (mon.2) 31 : audit [DBG] from='client.? 192.168.123.107:0/969698203' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:45.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[17657]: audit 2026-03-21T06:59:45.636588+0000 mon.b (mon.2) 31 : audit [DBG] from='client.? 192.168.123.107:0/969698203' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:44.882493+0000 mon.c (mon.1) 46 : audit [DBG] from='client.? 192.168.123.107:0/1856718767' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:44.882493+0000 mon.c (mon.1) 46 : audit [DBG] from='client.? 192.168.123.107:0/1856718767' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:44.903791+0000 mon.c (mon.1) 47 : audit [DBG] from='client.? 192.168.123.107:0/570228412' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:44.903791+0000 mon.c (mon.1) 47 : audit [DBG] from='client.? 192.168.123.107:0/570228412' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:44.911915+0000 mon.c (mon.1) 48 : audit [DBG] from='client.? 192.168.123.107:0/2452525968' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:44.911915+0000 mon.c (mon.1) 48 : audit [DBG] from='client.? 192.168.123.107:0/2452525968' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.019793+0000 mon.c (mon.1) 49 : audit [DBG] from='client.? 192.168.123.107:0/3069139021' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.019793+0000 mon.c (mon.1) 49 : audit [DBG] from='client.? 192.168.123.107:0/3069139021' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.026861+0000 mon.c (mon.1) 50 : audit [DBG] from='client.? 192.168.123.107:0/4165205942' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.026861+0000 mon.c (mon.1) 50 : audit [DBG] from='client.? 192.168.123.107:0/4165205942' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.473840+0000 mon.b (mon.2) 30 : audit [DBG] from='client.? 192.168.123.107:0/1347450667' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.473840+0000 mon.b (mon.2) 30 : audit [DBG] from='client.? 192.168.123.107:0/1347450667' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.501113+0000 mon.c (mon.1) 51 : audit [DBG] from='client.? 192.168.123.107:0/2220525325' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.501113+0000 mon.c (mon.1) 51 : audit [DBG] from='client.? 192.168.123.107:0/2220525325' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.510884+0000 mon.c (mon.1) 52 : audit [DBG] from='client.? 192.168.123.107:0/4276385491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.510884+0000 mon.c (mon.1) 52 : audit [DBG] from='client.? 192.168.123.107:0/4276385491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.628746+0000 mon.c (mon.1) 53 : audit [DBG] from='client.? 192.168.123.107:0/886655627' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.628746+0000 mon.c (mon.1) 53 : audit [DBG] from='client.? 192.168.123.107:0/886655627' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.636588+0000 mon.b (mon.2) 31 : audit [DBG] from='client.? 192.168.123.107:0/969698203' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:46.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:45 vm04 bash[20194]: audit 2026-03-21T06:59:45.636588+0000 mon.b (mon.2) 31 : audit [DBG] from='client.? 192.168.123.107:0/969698203' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:46.349 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:45 vm02 bash[49158]: debug Change detected - internal 8 / xattr 9 refreshing 2026-03-21T06:59:46.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.989 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.989 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.989 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.989 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: cluster 2026-03-21T06:59:45.644431+0000 mgr.x (mgr.14152) 516 : cluster [DBG] pgmap v407: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 85 KiB/s rd, 18 KiB/s wr, 60 op/s 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: cluster 2026-03-21T06:59:45.644431+0000 mgr.x (mgr.14152) 516 : cluster [DBG] pgmap v407: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 85 KiB/s rd, 18 KiB/s wr, 60 op/s 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: audit 2026-03-21T06:59:46.221640+0000 mon.c (mon.1) 54 : audit [DBG] from='client.? 192.168.123.107:0/3710482323' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: audit 2026-03-21T06:59:46.221640+0000 mon.c (mon.1) 54 : audit [DBG] from='client.? 192.168.123.107:0/3710482323' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: audit 2026-03-21T06:59:46.246210+0000 mon.c (mon.1) 55 : audit [DBG] from='client.? 192.168.123.107:0/3869462380' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: audit 2026-03-21T06:59:46.246210+0000 mon.c (mon.1) 55 : audit [DBG] from='client.? 192.168.123.107:0/3869462380' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: audit 2026-03-21T06:59:46.256572+0000 mon.c (mon.1) 56 : audit [DBG] from='client.? 192.168.123.107:0/3279648979' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: audit 2026-03-21T06:59:46.256572+0000 mon.c (mon.1) 56 : audit [DBG] from='client.? 192.168.123.107:0/3279648979' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: audit 2026-03-21T06:59:46.366055+0000 mon.c (mon.1) 57 : audit [DBG] from='client.? 192.168.123.107:0/1019372761' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: audit 2026-03-21T06:59:46.366055+0000 mon.c (mon.1) 57 : audit [DBG] from='client.? 192.168.123.107:0/1019372761' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: audit 2026-03-21T06:59:46.375326+0000 mon.c (mon.1) 58 : audit [DBG] from='client.? 192.168.123.107:0/62687453' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:46.989 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[17657]: audit 2026-03-21T06:59:46.375326+0000 mon.c (mon.1) 58 : audit [DBG] from='client.? 192.168.123.107:0/62687453' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:46.991 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.991 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.991 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:46.991 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:46.991 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.991 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.992 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.992 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:46] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:46.992 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.992 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.992 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.992 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: cluster 2026-03-21T06:59:45.644431+0000 mgr.x (mgr.14152) 516 : cluster [DBG] pgmap v407: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 85 KiB/s rd, 18 KiB/s wr, 60 op/s 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: cluster 2026-03-21T06:59:45.644431+0000 mgr.x (mgr.14152) 516 : cluster [DBG] pgmap v407: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 85 KiB/s rd, 18 KiB/s wr, 60 op/s 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: audit 2026-03-21T06:59:46.221640+0000 mon.c (mon.1) 54 : audit [DBG] from='client.? 192.168.123.107:0/3710482323' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: audit 2026-03-21T06:59:46.221640+0000 mon.c (mon.1) 54 : audit [DBG] from='client.? 192.168.123.107:0/3710482323' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: audit 2026-03-21T06:59:46.246210+0000 mon.c (mon.1) 55 : audit [DBG] from='client.? 192.168.123.107:0/3869462380' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: audit 2026-03-21T06:59:46.246210+0000 mon.c (mon.1) 55 : audit [DBG] from='client.? 192.168.123.107:0/3869462380' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: audit 2026-03-21T06:59:46.256572+0000 mon.c (mon.1) 56 : audit [DBG] from='client.? 192.168.123.107:0/3279648979' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: audit 2026-03-21T06:59:46.256572+0000 mon.c (mon.1) 56 : audit [DBG] from='client.? 192.168.123.107:0/3279648979' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: audit 2026-03-21T06:59:46.366055+0000 mon.c (mon.1) 57 : audit [DBG] from='client.? 192.168.123.107:0/1019372761' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: audit 2026-03-21T06:59:46.366055+0000 mon.c (mon.1) 57 : audit [DBG] from='client.? 192.168.123.107:0/1019372761' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: audit 2026-03-21T06:59:46.375326+0000 mon.c (mon.1) 58 : audit [DBG] from='client.? 192.168.123.107:0/62687453' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:46.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:46 vm07 bash[19945]: audit 2026-03-21T06:59:46.375326+0000 mon.c (mon.1) 58 : audit [DBG] from='client.? 192.168.123.107:0/62687453' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: cluster 2026-03-21T06:59:45.644431+0000 mgr.x (mgr.14152) 516 : cluster [DBG] pgmap v407: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 85 KiB/s rd, 18 KiB/s wr, 60 op/s 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: cluster 2026-03-21T06:59:45.644431+0000 mgr.x (mgr.14152) 516 : cluster [DBG] pgmap v407: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 85 KiB/s rd, 18 KiB/s wr, 60 op/s 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: audit 2026-03-21T06:59:46.221640+0000 mon.c (mon.1) 54 : audit [DBG] from='client.? 192.168.123.107:0/3710482323' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: audit 2026-03-21T06:59:46.221640+0000 mon.c (mon.1) 54 : audit [DBG] from='client.? 192.168.123.107:0/3710482323' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: audit 2026-03-21T06:59:46.246210+0000 mon.c (mon.1) 55 : audit [DBG] from='client.? 192.168.123.107:0/3869462380' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: audit 2026-03-21T06:59:46.246210+0000 mon.c (mon.1) 55 : audit [DBG] from='client.? 192.168.123.107:0/3869462380' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: audit 2026-03-21T06:59:46.256572+0000 mon.c (mon.1) 56 : audit [DBG] from='client.? 192.168.123.107:0/3279648979' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: audit 2026-03-21T06:59:46.256572+0000 mon.c (mon.1) 56 : audit [DBG] from='client.? 192.168.123.107:0/3279648979' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: audit 2026-03-21T06:59:46.366055+0000 mon.c (mon.1) 57 : audit [DBG] from='client.? 192.168.123.107:0/1019372761' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: audit 2026-03-21T06:59:46.366055+0000 mon.c (mon.1) 57 : audit [DBG] from='client.? 192.168.123.107:0/1019372761' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: audit 2026-03-21T06:59:46.375326+0000 mon.c (mon.1) 58 : audit [DBG] from='client.? 192.168.123.107:0/62687453' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:46 vm04 bash[20194]: audit 2026-03-21T06:59:46.375326+0000 mon.c (mon.1) 58 : audit [DBG] from='client.? 192.168.123.107:0/62687453' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:47.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:46 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[49158]: debug LUN unmap request received, config commit to be performed by vm07.local 2026-03-21T06:59:47.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:47.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:47.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:46 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:46] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug _targetlun update on vm02.local, successful 2026-03-21T06:59:47.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug LUN unmap request received, config commit to be performed by vm07.local 2026-03-21T06:59:47.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:47.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:47.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug _targetlun update on vm07.local, successful 2026-03-21T06:59:47.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:47] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:47.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:47] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:46.815357+0000 mon.c (mon.1) 59 : audit [DBG] from='client.? 192.168.123.107:0/3920105240' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:46.815357+0000 mon.c (mon.1) 59 : audit [DBG] from='client.? 192.168.123.107:0/3920105240' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:46.839525+0000 mon.c (mon.1) 60 : audit [DBG] from='client.? 192.168.123.107:0/1207491774' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:46.839525+0000 mon.c (mon.1) 60 : audit [DBG] from='client.? 192.168.123.107:0/1207491774' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:46.848836+0000 mon.c (mon.1) 61 : audit [DBG] from='client.? 192.168.123.107:0/307756951' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:46.848836+0000 mon.c (mon.1) 61 : audit [DBG] from='client.? 192.168.123.107:0/307756951' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:46.974023+0000 mon.a (mon.0) 997 : audit [DBG] from='client.? 192.168.123.107:0/3108909697' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:46.974023+0000 mon.a (mon.0) 997 : audit [DBG] from='client.? 192.168.123.107:0/3108909697' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:46.979049+0000 mon.b (mon.2) 32 : audit [DBG] from='client.? 192.168.123.107:0/2810787210' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:46.979049+0000 mon.b (mon.2) 32 : audit [DBG] from='client.? 192.168.123.107:0/2810787210' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:47.542824+0000 mon.c (mon.1) 62 : audit [DBG] from='client.? 192.168.123.107:0/929251515' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:47.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:47.542824+0000 mon.c (mon.1) 62 : audit [DBG] from='client.? 192.168.123.107:0/929251515' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:47.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:47.563629+0000 mon.c (mon.1) 63 : audit [DBG] from='client.? 192.168.123.107:0/514727204' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:47.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:47.563629+0000 mon.c (mon.1) 63 : audit [DBG] from='client.? 192.168.123.107:0/514727204' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:47.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:47.571943+0000 mon.c (mon.1) 64 : audit [DBG] from='client.? 192.168.123.107:0/4128115452' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:47.571943+0000 mon.c (mon.1) 64 : audit [DBG] from='client.? 192.168.123.107:0/4128115452' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:47.666714+0000 mon.c (mon.1) 65 : audit [DBG] from='client.? 192.168.123.107:0/190630205' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:47.666714+0000 mon.c (mon.1) 65 : audit [DBG] from='client.? 192.168.123.107:0/190630205' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:47.678929+0000 mon.a (mon.0) 998 : audit [DBG] from='client.? 192.168.123.107:0/1300783219' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:47.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:47 vm07 bash[19945]: audit 2026-03-21T06:59:47.678929+0000 mon.a (mon.0) 998 : audit [DBG] from='client.? 192.168.123.107:0/1300783219' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:47] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:47] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:47] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:47] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:47] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:47] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:47] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:47] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:47 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.991 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.991 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.991 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.991 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:47] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:46.815357+0000 mon.c (mon.1) 59 : audit [DBG] from='client.? 192.168.123.107:0/3920105240' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:46.815357+0000 mon.c (mon.1) 59 : audit [DBG] from='client.? 192.168.123.107:0/3920105240' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:46.839525+0000 mon.c (mon.1) 60 : audit [DBG] from='client.? 192.168.123.107:0/1207491774' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:46.839525+0000 mon.c (mon.1) 60 : audit [DBG] from='client.? 192.168.123.107:0/1207491774' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:46.848836+0000 mon.c (mon.1) 61 : audit [DBG] from='client.? 192.168.123.107:0/307756951' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:46.848836+0000 mon.c (mon.1) 61 : audit [DBG] from='client.? 192.168.123.107:0/307756951' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:46.974023+0000 mon.a (mon.0) 997 : audit [DBG] from='client.? 192.168.123.107:0/3108909697' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:46.974023+0000 mon.a (mon.0) 997 : audit [DBG] from='client.? 192.168.123.107:0/3108909697' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:46.979049+0000 mon.b (mon.2) 32 : audit [DBG] from='client.? 192.168.123.107:0/2810787210' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:46.979049+0000 mon.b (mon.2) 32 : audit [DBG] from='client.? 192.168.123.107:0/2810787210' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:47.542824+0000 mon.c (mon.1) 62 : audit [DBG] from='client.? 192.168.123.107:0/929251515' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:47.542824+0000 mon.c (mon.1) 62 : audit [DBG] from='client.? 192.168.123.107:0/929251515' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:47.563629+0000 mon.c (mon.1) 63 : audit [DBG] from='client.? 192.168.123.107:0/514727204' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:47.563629+0000 mon.c (mon.1) 63 : audit [DBG] from='client.? 192.168.123.107:0/514727204' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:47.571943+0000 mon.c (mon.1) 64 : audit [DBG] from='client.? 192.168.123.107:0/4128115452' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:47.571943+0000 mon.c (mon.1) 64 : audit [DBG] from='client.? 192.168.123.107:0/4128115452' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:47.666714+0000 mon.c (mon.1) 65 : audit [DBG] from='client.? 192.168.123.107:0/190630205' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:47.666714+0000 mon.c (mon.1) 65 : audit [DBG] from='client.? 192.168.123.107:0/190630205' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:47.678929+0000 mon.a (mon.0) 998 : audit [DBG] from='client.? 192.168.123.107:0/1300783219' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:47.992 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[17657]: audit 2026-03-21T06:59:47.678929+0000 mon.a (mon.0) 998 : audit [DBG] from='client.? 192.168.123.107:0/1300783219' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:46.815357+0000 mon.c (mon.1) 59 : audit [DBG] from='client.? 192.168.123.107:0/3920105240' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:46.815357+0000 mon.c (mon.1) 59 : audit [DBG] from='client.? 192.168.123.107:0/3920105240' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:46.839525+0000 mon.c (mon.1) 60 : audit [DBG] from='client.? 192.168.123.107:0/1207491774' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:46.839525+0000 mon.c (mon.1) 60 : audit [DBG] from='client.? 192.168.123.107:0/1207491774' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:46.848836+0000 mon.c (mon.1) 61 : audit [DBG] from='client.? 192.168.123.107:0/307756951' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:46.848836+0000 mon.c (mon.1) 61 : audit [DBG] from='client.? 192.168.123.107:0/307756951' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:46.974023+0000 mon.a (mon.0) 997 : audit [DBG] from='client.? 192.168.123.107:0/3108909697' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:46.974023+0000 mon.a (mon.0) 997 : audit [DBG] from='client.? 192.168.123.107:0/3108909697' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:46.979049+0000 mon.b (mon.2) 32 : audit [DBG] from='client.? 192.168.123.107:0/2810787210' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:46.979049+0000 mon.b (mon.2) 32 : audit [DBG] from='client.? 192.168.123.107:0/2810787210' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:47.542824+0000 mon.c (mon.1) 62 : audit [DBG] from='client.? 192.168.123.107:0/929251515' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:47.542824+0000 mon.c (mon.1) 62 : audit [DBG] from='client.? 192.168.123.107:0/929251515' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:47.563629+0000 mon.c (mon.1) 63 : audit [DBG] from='client.? 192.168.123.107:0/514727204' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:47.563629+0000 mon.c (mon.1) 63 : audit [DBG] from='client.? 192.168.123.107:0/514727204' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:47.571943+0000 mon.c (mon.1) 64 : audit [DBG] from='client.? 192.168.123.107:0/4128115452' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:47.571943+0000 mon.c (mon.1) 64 : audit [DBG] from='client.? 192.168.123.107:0/4128115452' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:47.666714+0000 mon.c (mon.1) 65 : audit [DBG] from='client.? 192.168.123.107:0/190630205' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:47.666714+0000 mon.c (mon.1) 65 : audit [DBG] from='client.? 192.168.123.107:0/190630205' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:47.678929+0000 mon.a (mon.0) 998 : audit [DBG] from='client.? 192.168.123.107:0/1300783219' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:48.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:47 vm04 bash[20194]: audit 2026-03-21T06:59:47.678929+0000 mon.a (mon.0) 998 : audit [DBG] from='client.? 192.168.123.107:0/1300783219' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:48.290 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:47 vm02 bash[49158]: debug Change detected - internal 9 / xattr 10 refreshing 2026-03-21T06:59:48.290 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.290 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.290 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.290 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.290 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.290 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.290 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[49158]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T06:59:48.290 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[49158]: debug Removing target configuration 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug _target update on vm02.local, successful 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug Removing target configuration 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug _target update on vm07.local, successful 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:48.379 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:48.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:48.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:48] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T06:59:49.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[49158]: debug Change detected - internal 11 / xattr 12 refreshing 2026-03-21T06:59:49.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:46.820485+0000 mgr.x (mgr.14152) 517 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:46.820485+0000 mgr.x (mgr.14152) 517 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: cluster 2026-03-21T06:59:47.644701+0000 mgr.x (mgr.14152) 518 : cluster [DBG] pgmap v408: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 40 KiB/s rd, 17 KiB/s wr, 14 op/s 2026-03-21T06:59:49.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: cluster 2026-03-21T06:59:47.644701+0000 mgr.x (mgr.14152) 518 : cluster [DBG] pgmap v408: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 40 KiB/s rd, 17 KiB/s wr, 14 op/s 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:47.653987+0000 mgr.x (mgr.14152) 519 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:47.653987+0000 mgr.x (mgr.14152) 519 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.079394+0000 mon.c (mon.1) 66 : audit [DBG] from='client.? 192.168.123.107:0/375898732' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.079394+0000 mon.c (mon.1) 66 : audit [DBG] from='client.? 192.168.123.107:0/375898732' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.099820+0000 mon.c (mon.1) 67 : audit [DBG] from='client.? 192.168.123.107:0/4051559609' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.099820+0000 mon.c (mon.1) 67 : audit [DBG] from='client.? 192.168.123.107:0/4051559609' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.107253+0000 mon.c (mon.1) 68 : audit [DBG] from='client.? 192.168.123.107:0/1543795166' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.107253+0000 mon.c (mon.1) 68 : audit [DBG] from='client.? 192.168.123.107:0/1543795166' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.201020+0000 mon.c (mon.1) 69 : audit [DBG] from='client.? 192.168.123.107:0/3465068260' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.201020+0000 mon.c (mon.1) 69 : audit [DBG] from='client.? 192.168.123.107:0/3465068260' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.208061+0000 mon.c (mon.1) 70 : audit [DBG] from='client.? 192.168.123.107:0/532957146' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.208061+0000 mon.c (mon.1) 70 : audit [DBG] from='client.? 192.168.123.107:0/532957146' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.759022+0000 mon.c (mon.1) 71 : audit [DBG] from='client.? 192.168.123.107:0/1224759060' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:48 vm02 bash[17657]: audit 2026-03-21T06:59:48.759022+0000 mon.c (mon.1) 71 : audit [DBG] from='client.? 192.168.123.107:0/1224759060' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:49.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:49.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:49.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:48 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:48] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:46.820485+0000 mgr.x (mgr.14152) 517 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:46.820485+0000 mgr.x (mgr.14152) 517 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: cluster 2026-03-21T06:59:47.644701+0000 mgr.x (mgr.14152) 518 : cluster [DBG] pgmap v408: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 40 KiB/s rd, 17 KiB/s wr, 14 op/s 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: cluster 2026-03-21T06:59:47.644701+0000 mgr.x (mgr.14152) 518 : cluster [DBG] pgmap v408: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 40 KiB/s rd, 17 KiB/s wr, 14 op/s 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:47.653987+0000 mgr.x (mgr.14152) 519 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:47.653987+0000 mgr.x (mgr.14152) 519 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.079394+0000 mon.c (mon.1) 66 : audit [DBG] from='client.? 192.168.123.107:0/375898732' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.079394+0000 mon.c (mon.1) 66 : audit [DBG] from='client.? 192.168.123.107:0/375898732' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.099820+0000 mon.c (mon.1) 67 : audit [DBG] from='client.? 192.168.123.107:0/4051559609' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.099820+0000 mon.c (mon.1) 67 : audit [DBG] from='client.? 192.168.123.107:0/4051559609' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.107253+0000 mon.c (mon.1) 68 : audit [DBG] from='client.? 192.168.123.107:0/1543795166' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.107253+0000 mon.c (mon.1) 68 : audit [DBG] from='client.? 192.168.123.107:0/1543795166' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.201020+0000 mon.c (mon.1) 69 : audit [DBG] from='client.? 192.168.123.107:0/3465068260' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.201020+0000 mon.c (mon.1) 69 : audit [DBG] from='client.? 192.168.123.107:0/3465068260' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.208061+0000 mon.c (mon.1) 70 : audit [DBG] from='client.? 192.168.123.107:0/532957146' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:49.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.208061+0000 mon.c (mon.1) 70 : audit [DBG] from='client.? 192.168.123.107:0/532957146' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:49.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.759022+0000 mon.c (mon.1) 71 : audit [DBG] from='client.? 192.168.123.107:0/1224759060' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:48 vm07 bash[19945]: audit 2026-03-21T06:59:48.759022+0000 mon.c (mon.1) 71 : audit [DBG] from='client.? 192.168.123.107:0/1224759060' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:46.820485+0000 mgr.x (mgr.14152) 517 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:46.820485+0000 mgr.x (mgr.14152) 517 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: cluster 2026-03-21T06:59:47.644701+0000 mgr.x (mgr.14152) 518 : cluster [DBG] pgmap v408: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 40 KiB/s rd, 17 KiB/s wr, 14 op/s 2026-03-21T06:59:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: cluster 2026-03-21T06:59:47.644701+0000 mgr.x (mgr.14152) 518 : cluster [DBG] pgmap v408: 4 pgs: 4 active+clean; 8.5 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 40 KiB/s rd, 17 KiB/s wr, 14 op/s 2026-03-21T06:59:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:47.653987+0000 mgr.x (mgr.14152) 519 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:47.653987+0000 mgr.x (mgr.14152) 519 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.079394+0000 mon.c (mon.1) 66 : audit [DBG] from='client.? 192.168.123.107:0/375898732' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.079394+0000 mon.c (mon.1) 66 : audit [DBG] from='client.? 192.168.123.107:0/375898732' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.099820+0000 mon.c (mon.1) 67 : audit [DBG] from='client.? 192.168.123.107:0/4051559609' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:49.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.099820+0000 mon.c (mon.1) 67 : audit [DBG] from='client.? 192.168.123.107:0/4051559609' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:49.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.107253+0000 mon.c (mon.1) 68 : audit [DBG] from='client.? 192.168.123.107:0/1543795166' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.107253+0000 mon.c (mon.1) 68 : audit [DBG] from='client.? 192.168.123.107:0/1543795166' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.201020+0000 mon.c (mon.1) 69 : audit [DBG] from='client.? 192.168.123.107:0/3465068260' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.201020+0000 mon.c (mon.1) 69 : audit [DBG] from='client.? 192.168.123.107:0/3465068260' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.208061+0000 mon.c (mon.1) 70 : audit [DBG] from='client.? 192.168.123.107:0/532957146' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:49.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.208061+0000 mon.c (mon.1) 70 : audit [DBG] from='client.? 192.168.123.107:0/532957146' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:49.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.759022+0000 mon.c (mon.1) 71 : audit [DBG] from='client.? 192.168.123.107:0/1224759060' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:48 vm04 bash[20194]: audit 2026-03-21T06:59:48.759022+0000 mon.c (mon.1) 71 : audit [DBG] from='client.? 192.168.123.107:0/1224759060' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:49] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:49] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:49] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:49] "GET /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:49] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:49] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: debug LUN deletion request received, rbd removal to be performed by vm07.local 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:49] "DELETE /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:49] "DELETE /api/_disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: debug _disk update on vm07.local, successful 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:49] "DELETE /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:49.586 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:49 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 06:59:49] "DELETE /api/disk/datapool/block0 HTTP/1.1" 200 - 2026-03-21T06:59:49.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:48.780324+0000 mon.c (mon.1) 72 : audit [DBG] from='client.? 192.168.123.107:0/1764178393' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:49.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:48.780324+0000 mon.c (mon.1) 72 : audit [DBG] from='client.? 192.168.123.107:0/1764178393' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:49.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:48.788541+0000 mon.c (mon.1) 73 : audit [DBG] from='client.? 192.168.123.107:0/3790405031' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:48.788541+0000 mon.c (mon.1) 73 : audit [DBG] from='client.? 192.168.123.107:0/3790405031' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:48.860620+0000 mon.c (mon.1) 74 : audit [DBG] from='client.? 192.168.123.107:0/503001345' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:48.860620+0000 mon.c (mon.1) 74 : audit [DBG] from='client.? 192.168.123.107:0/503001345' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:48.867952+0000 mon.c (mon.1) 75 : audit [DBG] from='client.? 192.168.123.107:0/3630150925' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:48.867952+0000 mon.c (mon.1) 75 : audit [DBG] from='client.? 192.168.123.107:0/3630150925' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.284940+0000 mon.c (mon.1) 76 : audit [DBG] from='client.? 192.168.123.107:0/3897053893' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.284940+0000 mon.c (mon.1) 76 : audit [DBG] from='client.? 192.168.123.107:0/3897053893' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.309344+0000 mon.c (mon.1) 77 : audit [DBG] from='client.? 192.168.123.107:0/3040594968' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.309344+0000 mon.c (mon.1) 77 : audit [DBG] from='client.? 192.168.123.107:0/3040594968' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.321166+0000 mon.c (mon.1) 78 : audit [DBG] from='client.? 192.168.123.107:0/1888893491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.321166+0000 mon.c (mon.1) 78 : audit [DBG] from='client.? 192.168.123.107:0/1888893491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.400081+0000 mon.b (mon.2) 33 : audit [DBG] from='client.? 192.168.123.107:0/2246620908' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.400081+0000 mon.b (mon.2) 33 : audit [DBG] from='client.? 192.168.123.107:0/2246620908' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.410238+0000 mon.c (mon.1) 79 : audit [DBG] from='client.? 192.168.123.107:0/1543971456' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.410238+0000 mon.c (mon.1) 79 : audit [DBG] from='client.? 192.168.123.107:0/1543971456' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.554163+0000 mon.c (mon.1) 80 : audit [DBG] from='client.? 192.168.123.107:0/1404729917' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:49.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:49 vm07 bash[19945]: audit 2026-03-21T06:59:49.554163+0000 mon.c (mon.1) 80 : audit [DBG] from='client.? 192.168.123.107:0/1404729917' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[49158]: debug Change detected - internal 12 / xattr 13 refreshing 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:48.780324+0000 mon.c (mon.1) 72 : audit [DBG] from='client.? 192.168.123.107:0/1764178393' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:48.780324+0000 mon.c (mon.1) 72 : audit [DBG] from='client.? 192.168.123.107:0/1764178393' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:48.788541+0000 mon.c (mon.1) 73 : audit [DBG] from='client.? 192.168.123.107:0/3790405031' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:48.788541+0000 mon.c (mon.1) 73 : audit [DBG] from='client.? 192.168.123.107:0/3790405031' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:48.860620+0000 mon.c (mon.1) 74 : audit [DBG] from='client.? 192.168.123.107:0/503001345' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:48.860620+0000 mon.c (mon.1) 74 : audit [DBG] from='client.? 192.168.123.107:0/503001345' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:48.867952+0000 mon.c (mon.1) 75 : audit [DBG] from='client.? 192.168.123.107:0/3630150925' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:48.867952+0000 mon.c (mon.1) 75 : audit [DBG] from='client.? 192.168.123.107:0/3630150925' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.284940+0000 mon.c (mon.1) 76 : audit [DBG] from='client.? 192.168.123.107:0/3897053893' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.284940+0000 mon.c (mon.1) 76 : audit [DBG] from='client.? 192.168.123.107:0/3897053893' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.309344+0000 mon.c (mon.1) 77 : audit [DBG] from='client.? 192.168.123.107:0/3040594968' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.309344+0000 mon.c (mon.1) 77 : audit [DBG] from='client.? 192.168.123.107:0/3040594968' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.321166+0000 mon.c (mon.1) 78 : audit [DBG] from='client.? 192.168.123.107:0/1888893491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.321166+0000 mon.c (mon.1) 78 : audit [DBG] from='client.? 192.168.123.107:0/1888893491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.400081+0000 mon.b (mon.2) 33 : audit [DBG] from='client.? 192.168.123.107:0/2246620908' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.400081+0000 mon.b (mon.2) 33 : audit [DBG] from='client.? 192.168.123.107:0/2246620908' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.410238+0000 mon.c (mon.1) 79 : audit [DBG] from='client.? 192.168.123.107:0/1543971456' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.410238+0000 mon.c (mon.1) 79 : audit [DBG] from='client.? 192.168.123.107:0/1543971456' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.554163+0000 mon.c (mon.1) 80 : audit [DBG] from='client.? 192.168.123.107:0/1404729917' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:49 vm02 bash[17657]: audit 2026-03-21T06:59:49.554163+0000 mon.c (mon.1) 80 : audit [DBG] from='client.? 192.168.123.107:0/1404729917' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.147 INFO:tasks.cram.client.2.vm07.stdout:/home/ubuntu/cephtest/archive/cram.client.2/gwcli_delete.t: passed 2026-03-21T06:59:50.147 INFO:tasks.cram.client.2.vm07.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-21T06:59:50.153 DEBUG:teuthology.orchestra.run.vm02:> test -f /home/ubuntu/cephtest/archive/cram.client.0/gwcli_create.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.0/gwcli_create.t 2026-03-21T06:59:50.158 DEBUG:teuthology.orchestra.run.vm02:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.0 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.0 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:48.780324+0000 mon.c (mon.1) 72 : audit [DBG] from='client.? 192.168.123.107:0/1764178393' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:48.780324+0000 mon.c (mon.1) 72 : audit [DBG] from='client.? 192.168.123.107:0/1764178393' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:48.788541+0000 mon.c (mon.1) 73 : audit [DBG] from='client.? 192.168.123.107:0/3790405031' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:48.788541+0000 mon.c (mon.1) 73 : audit [DBG] from='client.? 192.168.123.107:0/3790405031' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:48.860620+0000 mon.c (mon.1) 74 : audit [DBG] from='client.? 192.168.123.107:0/503001345' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:48.860620+0000 mon.c (mon.1) 74 : audit [DBG] from='client.? 192.168.123.107:0/503001345' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:48.867952+0000 mon.c (mon.1) 75 : audit [DBG] from='client.? 192.168.123.107:0/3630150925' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:48.867952+0000 mon.c (mon.1) 75 : audit [DBG] from='client.? 192.168.123.107:0/3630150925' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.284940+0000 mon.c (mon.1) 76 : audit [DBG] from='client.? 192.168.123.107:0/3897053893' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.284940+0000 mon.c (mon.1) 76 : audit [DBG] from='client.? 192.168.123.107:0/3897053893' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.309344+0000 mon.c (mon.1) 77 : audit [DBG] from='client.? 192.168.123.107:0/3040594968' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.309344+0000 mon.c (mon.1) 77 : audit [DBG] from='client.? 192.168.123.107:0/3040594968' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.321166+0000 mon.c (mon.1) 78 : audit [DBG] from='client.? 192.168.123.107:0/1888893491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.321166+0000 mon.c (mon.1) 78 : audit [DBG] from='client.? 192.168.123.107:0/1888893491' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.400081+0000 mon.b (mon.2) 33 : audit [DBG] from='client.? 192.168.123.107:0/2246620908' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.400081+0000 mon.b (mon.2) 33 : audit [DBG] from='client.? 192.168.123.107:0/2246620908' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.410238+0000 mon.c (mon.1) 79 : audit [DBG] from='client.? 192.168.123.107:0/1543971456' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.410238+0000 mon.c (mon.1) 79 : audit [DBG] from='client.? 192.168.123.107:0/1543971456' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.554163+0000 mon.c (mon.1) 80 : audit [DBG] from='client.? 192.168.123.107:0/1404729917' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:49 vm04 bash[20194]: audit 2026-03-21T06:59:49.554163+0000 mon.c (mon.1) 80 : audit [DBG] from='client.? 192.168.123.107:0/1404729917' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:50.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:50 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:50] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:50.402 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:50 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 06:59:50] "GET /api/config HTTP/1.1" 200 - 2026-03-21T06:59:50.616 DEBUG:teuthology.orchestra.run.vm04:> test -f /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t 2026-03-21T06:59:50.620 DEBUG:teuthology.orchestra.run.vm04:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.1 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.1 2026-03-21T06:59:51.069 DEBUG:teuthology.orchestra.run.vm07:> test -f /home/ubuntu/cephtest/archive/cram.client.2/gwcli_delete.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.2/gwcli_delete.t 2026-03-21T06:59:51.073 DEBUG:teuthology.orchestra.run.vm07:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.2 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.2 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: cluster 2026-03-21T06:59:49.645186+0000 mgr.x (mgr.14152) 520 : cluster [DBG] pgmap v409: 4 pgs: 4 active+clean; 11 MiB data, 220 MiB used, 160 GiB / 160 GiB avail; 65 KiB/s rd, 38 KiB/s wr, 33 op/s 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: cluster 2026-03-21T06:59:49.645186+0000 mgr.x (mgr.14152) 520 : cluster [DBG] pgmap v409: 4 pgs: 4 active+clean; 11 MiB data, 220 MiB used, 160 GiB / 160 GiB avail; 65 KiB/s rd, 38 KiB/s wr, 33 op/s 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: audit 2026-03-21T06:59:50.011328+0000 mon.c (mon.1) 81 : audit [DBG] from='client.? 192.168.123.107:0/308378403' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: audit 2026-03-21T06:59:50.011328+0000 mon.c (mon.1) 81 : audit [DBG] from='client.? 192.168.123.107:0/308378403' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: audit 2026-03-21T06:59:50.034800+0000 mon.c (mon.1) 82 : audit [DBG] from='client.? 192.168.123.107:0/172141789' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: audit 2026-03-21T06:59:50.034800+0000 mon.c (mon.1) 82 : audit [DBG] from='client.? 192.168.123.107:0/172141789' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: audit 2026-03-21T06:59:50.043503+0000 mon.c (mon.1) 83 : audit [DBG] from='client.? 192.168.123.107:0/3733930106' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: audit 2026-03-21T06:59:50.043503+0000 mon.c (mon.1) 83 : audit [DBG] from='client.? 192.168.123.107:0/3733930106' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: audit 2026-03-21T06:59:50.088907+0000 mon.c (mon.1) 84 : audit [DBG] from='client.? 192.168.123.107:0/2623112468' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: audit 2026-03-21T06:59:50.088907+0000 mon.c (mon.1) 84 : audit [DBG] from='client.? 192.168.123.107:0/2623112468' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: audit 2026-03-21T06:59:50.098896+0000 mon.c (mon.1) 85 : audit [DBG] from='client.? 192.168.123.107:0/231143603' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:51.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:50 vm02 bash[17657]: audit 2026-03-21T06:59:50.098896+0000 mon.c (mon.1) 85 : audit [DBG] from='client.? 192.168.123.107:0/231143603' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: cluster 2026-03-21T06:59:49.645186+0000 mgr.x (mgr.14152) 520 : cluster [DBG] pgmap v409: 4 pgs: 4 active+clean; 11 MiB data, 220 MiB used, 160 GiB / 160 GiB avail; 65 KiB/s rd, 38 KiB/s wr, 33 op/s 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: cluster 2026-03-21T06:59:49.645186+0000 mgr.x (mgr.14152) 520 : cluster [DBG] pgmap v409: 4 pgs: 4 active+clean; 11 MiB data, 220 MiB used, 160 GiB / 160 GiB avail; 65 KiB/s rd, 38 KiB/s wr, 33 op/s 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: audit 2026-03-21T06:59:50.011328+0000 mon.c (mon.1) 81 : audit [DBG] from='client.? 192.168.123.107:0/308378403' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: audit 2026-03-21T06:59:50.011328+0000 mon.c (mon.1) 81 : audit [DBG] from='client.? 192.168.123.107:0/308378403' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: audit 2026-03-21T06:59:50.034800+0000 mon.c (mon.1) 82 : audit [DBG] from='client.? 192.168.123.107:0/172141789' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: audit 2026-03-21T06:59:50.034800+0000 mon.c (mon.1) 82 : audit [DBG] from='client.? 192.168.123.107:0/172141789' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: audit 2026-03-21T06:59:50.043503+0000 mon.c (mon.1) 83 : audit [DBG] from='client.? 192.168.123.107:0/3733930106' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: audit 2026-03-21T06:59:50.043503+0000 mon.c (mon.1) 83 : audit [DBG] from='client.? 192.168.123.107:0/3733930106' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: audit 2026-03-21T06:59:50.088907+0000 mon.c (mon.1) 84 : audit [DBG] from='client.? 192.168.123.107:0/2623112468' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: audit 2026-03-21T06:59:50.088907+0000 mon.c (mon.1) 84 : audit [DBG] from='client.? 192.168.123.107:0/2623112468' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: audit 2026-03-21T06:59:50.098896+0000 mon.c (mon.1) 85 : audit [DBG] from='client.? 192.168.123.107:0/231143603' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:50 vm07 bash[19945]: audit 2026-03-21T06:59:50.098896+0000 mon.c (mon.1) 85 : audit [DBG] from='client.? 192.168.123.107:0/231143603' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: cluster 2026-03-21T06:59:49.645186+0000 mgr.x (mgr.14152) 520 : cluster [DBG] pgmap v409: 4 pgs: 4 active+clean; 11 MiB data, 220 MiB used, 160 GiB / 160 GiB avail; 65 KiB/s rd, 38 KiB/s wr, 33 op/s 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: cluster 2026-03-21T06:59:49.645186+0000 mgr.x (mgr.14152) 520 : cluster [DBG] pgmap v409: 4 pgs: 4 active+clean; 11 MiB data, 220 MiB used, 160 GiB / 160 GiB avail; 65 KiB/s rd, 38 KiB/s wr, 33 op/s 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: audit 2026-03-21T06:59:50.011328+0000 mon.c (mon.1) 81 : audit [DBG] from='client.? 192.168.123.107:0/308378403' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: audit 2026-03-21T06:59:50.011328+0000 mon.c (mon.1) 81 : audit [DBG] from='client.? 192.168.123.107:0/308378403' entity='client.iscsi.iscsi.b' cmd={"prefix": "version"} : dispatch 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: audit 2026-03-21T06:59:50.034800+0000 mon.c (mon.1) 82 : audit [DBG] from='client.? 192.168.123.107:0/172141789' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: audit 2026-03-21T06:59:50.034800+0000 mon.c (mon.1) 82 : audit [DBG] from='client.? 192.168.123.107:0/172141789' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd dump", "format": "json"} : dispatch 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: audit 2026-03-21T06:59:50.043503+0000 mon.c (mon.1) 83 : audit [DBG] from='client.? 192.168.123.107:0/3733930106' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: audit 2026-03-21T06:59:50.043503+0000 mon.c (mon.1) 83 : audit [DBG] from='client.? 192.168.123.107:0/3733930106' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: audit 2026-03-21T06:59:50.088907+0000 mon.c (mon.1) 84 : audit [DBG] from='client.? 192.168.123.107:0/2623112468' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: audit 2026-03-21T06:59:50.088907+0000 mon.c (mon.1) 84 : audit [DBG] from='client.? 192.168.123.107:0/2623112468' entity='client.iscsi.iscsi.b' cmd={"prefix": "status", "format": "json"} : dispatch 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: audit 2026-03-21T06:59:50.098896+0000 mon.c (mon.1) 85 : audit [DBG] from='client.? 192.168.123.107:0/231143603' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:51.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:50 vm04 bash[20194]: audit 2026-03-21T06:59:50.098896+0000 mon.c (mon.1) 85 : audit [DBG] from='client.? 192.168.123.107:0/231143603' entity='client.iscsi.iscsi.b' cmd={"prefix": "df", "format": "json"} : dispatch 2026-03-21T06:59:51.475 INFO:teuthology.run_tasks:Running task cram... 2026-03-21T06:59:51.477 INFO:tasks.cram:Pulling tests from https://github.com/kshtsk/ceph.git ref 0392f78529848ec72469e8e431875cb98d3a5fb4 2026-03-21T06:59:51.477 DEBUG:teuthology.orchestra.run.vm02:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.0 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-21T06:59:52.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:51 vm02 bash[17657]: cluster 2026-03-21T06:59:51.645545+0000 mgr.x (mgr.14152) 521 : cluster [DBG] pgmap v410: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 81 KiB/s rd, 39 KiB/s wr, 49 op/s 2026-03-21T06:59:52.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:51 vm02 bash[17657]: cluster 2026-03-21T06:59:51.645545+0000 mgr.x (mgr.14152) 521 : cluster [DBG] pgmap v410: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 81 KiB/s rd, 39 KiB/s wr, 49 op/s 2026-03-21T06:59:52.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:51 vm07 bash[19945]: cluster 2026-03-21T06:59:51.645545+0000 mgr.x (mgr.14152) 521 : cluster [DBG] pgmap v410: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 81 KiB/s rd, 39 KiB/s wr, 49 op/s 2026-03-21T06:59:52.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:51 vm07 bash[19945]: cluster 2026-03-21T06:59:51.645545+0000 mgr.x (mgr.14152) 521 : cluster [DBG] pgmap v410: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 81 KiB/s rd, 39 KiB/s wr, 49 op/s 2026-03-21T06:59:52.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:51 vm04 bash[20194]: cluster 2026-03-21T06:59:51.645545+0000 mgr.x (mgr.14152) 521 : cluster [DBG] pgmap v410: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 81 KiB/s rd, 39 KiB/s wr, 49 op/s 2026-03-21T06:59:52.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:51 vm04 bash[20194]: cluster 2026-03-21T06:59:51.645545+0000 mgr.x (mgr.14152) 521 : cluster [DBG] pgmap v410: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 81 KiB/s rd, 39 KiB/s wr, 49 op/s 2026-03-21T06:59:53.103 INFO:teuthology.orchestra.run.vm02.stdout:Collecting cram==0.6 2026-03-21T06:59:53.103 INFO:teuthology.orchestra.run.vm02.stdout: Using cached cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-21T06:59:53.113 INFO:teuthology.orchestra.run.vm02.stdout:Installing collected packages: cram 2026-03-21T06:59:53.119 INFO:teuthology.orchestra.run.vm02.stdout:Successfully installed cram-0.6 2026-03-21T06:59:53.145 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/clone.client.0 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.0 && cd /home/ubuntu/cephtest/clone.client.0 && git checkout 0392f78529848ec72469e8e431875cb98d3a5fb4 2026-03-21T06:59:53.149 INFO:teuthology.orchestra.run.vm02.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.0'... 2026-03-21T06:59:55.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:54 vm04 bash[20194]: cluster 2026-03-21T06:59:53.645833+0000 mgr.x (mgr.14152) 522 : cluster [DBG] pgmap v411: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 80 KiB/s rd, 39 KiB/s wr, 48 op/s 2026-03-21T06:59:55.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:54 vm04 bash[20194]: cluster 2026-03-21T06:59:53.645833+0000 mgr.x (mgr.14152) 522 : cluster [DBG] pgmap v411: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 80 KiB/s rd, 39 KiB/s wr, 48 op/s 2026-03-21T06:59:55.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:54 vm02 bash[17657]: cluster 2026-03-21T06:59:53.645833+0000 mgr.x (mgr.14152) 522 : cluster [DBG] pgmap v411: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 80 KiB/s rd, 39 KiB/s wr, 48 op/s 2026-03-21T06:59:55.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:54 vm02 bash[17657]: cluster 2026-03-21T06:59:53.645833+0000 mgr.x (mgr.14152) 522 : cluster [DBG] pgmap v411: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 80 KiB/s rd, 39 KiB/s wr, 48 op/s 2026-03-21T06:59:55.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:54 vm07 bash[19945]: cluster 2026-03-21T06:59:53.645833+0000 mgr.x (mgr.14152) 522 : cluster [DBG] pgmap v411: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 80 KiB/s rd, 39 KiB/s wr, 48 op/s 2026-03-21T06:59:55.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:54 vm07 bash[19945]: cluster 2026-03-21T06:59:53.645833+0000 mgr.x (mgr.14152) 522 : cluster [DBG] pgmap v411: 4 pgs: 4 active+clean; 2.6 MiB data, 219 MiB used, 160 GiB / 160 GiB avail; 80 KiB/s rd, 39 KiB/s wr, 48 op/s 2026-03-21T06:59:57.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:56 vm04 bash[20194]: cluster 2026-03-21T06:59:55.646137+0000 mgr.x (mgr.14152) 523 : cluster [DBG] pgmap v412: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 91 KiB/s rd, 39 KiB/s wr, 62 op/s 2026-03-21T06:59:57.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:56 vm04 bash[20194]: cluster 2026-03-21T06:59:55.646137+0000 mgr.x (mgr.14152) 523 : cluster [DBG] pgmap v412: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 91 KiB/s rd, 39 KiB/s wr, 62 op/s 2026-03-21T06:59:57.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 06:59:56 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T06:59:57.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:56 vm02 bash[17657]: cluster 2026-03-21T06:59:55.646137+0000 mgr.x (mgr.14152) 523 : cluster [DBG] pgmap v412: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 91 KiB/s rd, 39 KiB/s wr, 62 op/s 2026-03-21T06:59:57.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:56 vm02 bash[17657]: cluster 2026-03-21T06:59:55.646137+0000 mgr.x (mgr.14152) 523 : cluster [DBG] pgmap v412: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 91 KiB/s rd, 39 KiB/s wr, 62 op/s 2026-03-21T06:59:57.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:56 vm07 bash[19945]: cluster 2026-03-21T06:59:55.646137+0000 mgr.x (mgr.14152) 523 : cluster [DBG] pgmap v412: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 91 KiB/s rd, 39 KiB/s wr, 62 op/s 2026-03-21T06:59:57.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:56 vm07 bash[19945]: cluster 2026-03-21T06:59:55.646137+0000 mgr.x (mgr.14152) 523 : cluster [DBG] pgmap v412: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 91 KiB/s rd, 39 KiB/s wr, 62 op/s 2026-03-21T06:59:57.902 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 06:59:57 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T06:59:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:58 vm04 bash[20194]: audit 2026-03-21T06:59:56.824764+0000 mgr.x (mgr.14152) 524 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:58 vm04 bash[20194]: audit 2026-03-21T06:59:56.824764+0000 mgr.x (mgr.14152) 524 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:58 vm04 bash[20194]: cluster 2026-03-21T06:59:57.646469+0000 mgr.x (mgr.14152) 525 : cluster [DBG] pgmap v413: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T06:59:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:58 vm04 bash[20194]: cluster 2026-03-21T06:59:57.646469+0000 mgr.x (mgr.14152) 525 : cluster [DBG] pgmap v413: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T06:59:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:58 vm04 bash[20194]: audit 2026-03-21T06:59:57.655712+0000 mgr.x (mgr.14152) 526 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 06:59:58 vm04 bash[20194]: audit 2026-03-21T06:59:57.655712+0000 mgr.x (mgr.14152) 526 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:58 vm02 bash[17657]: audit 2026-03-21T06:59:56.824764+0000 mgr.x (mgr.14152) 524 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:58 vm02 bash[17657]: audit 2026-03-21T06:59:56.824764+0000 mgr.x (mgr.14152) 524 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:58 vm02 bash[17657]: cluster 2026-03-21T06:59:57.646469+0000 mgr.x (mgr.14152) 525 : cluster [DBG] pgmap v413: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T06:59:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:58 vm02 bash[17657]: cluster 2026-03-21T06:59:57.646469+0000 mgr.x (mgr.14152) 525 : cluster [DBG] pgmap v413: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T06:59:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:58 vm02 bash[17657]: audit 2026-03-21T06:59:57.655712+0000 mgr.x (mgr.14152) 526 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 06:59:58 vm02 bash[17657]: audit 2026-03-21T06:59:57.655712+0000 mgr.x (mgr.14152) 526 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:58 vm07 bash[19945]: audit 2026-03-21T06:59:56.824764+0000 mgr.x (mgr.14152) 524 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:58 vm07 bash[19945]: audit 2026-03-21T06:59:56.824764+0000 mgr.x (mgr.14152) 524 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:58 vm07 bash[19945]: cluster 2026-03-21T06:59:57.646469+0000 mgr.x (mgr.14152) 525 : cluster [DBG] pgmap v413: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T06:59:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:58 vm07 bash[19945]: cluster 2026-03-21T06:59:57.646469+0000 mgr.x (mgr.14152) 525 : cluster [DBG] pgmap v413: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T06:59:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:58 vm07 bash[19945]: audit 2026-03-21T06:59:57.655712+0000 mgr.x (mgr.14152) 526 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T06:59:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 06:59:58 vm07 bash[19945]: audit 2026-03-21T06:59:57.655712+0000 mgr.x (mgr.14152) 526 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:01.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:00 vm02 bash[17657]: cluster 2026-03-21T06:59:59.646801+0000 mgr.x (mgr.14152) 527 : cluster [DBG] pgmap v414: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T07:00:01.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:00 vm02 bash[17657]: cluster 2026-03-21T06:59:59.646801+0000 mgr.x (mgr.14152) 527 : cluster [DBG] pgmap v414: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T07:00:01.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:00 vm02 bash[17657]: cluster 2026-03-21T07:00:00.000133+0000 mon.a (mon.0) 999 : cluster [INF] overall HEALTH_OK 2026-03-21T07:00:01.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:00 vm02 bash[17657]: cluster 2026-03-21T07:00:00.000133+0000 mon.a (mon.0) 999 : cluster [INF] overall HEALTH_OK 2026-03-21T07:00:01.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:00 vm07 bash[19945]: cluster 2026-03-21T06:59:59.646801+0000 mgr.x (mgr.14152) 527 : cluster [DBG] pgmap v414: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T07:00:01.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:00 vm07 bash[19945]: cluster 2026-03-21T06:59:59.646801+0000 mgr.x (mgr.14152) 527 : cluster [DBG] pgmap v414: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T07:00:01.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:00 vm07 bash[19945]: cluster 2026-03-21T07:00:00.000133+0000 mon.a (mon.0) 999 : cluster [INF] overall HEALTH_OK 2026-03-21T07:00:01.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:00 vm07 bash[19945]: cluster 2026-03-21T07:00:00.000133+0000 mon.a (mon.0) 999 : cluster [INF] overall HEALTH_OK 2026-03-21T07:00:01.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:00 vm04 bash[20194]: cluster 2026-03-21T06:59:59.646801+0000 mgr.x (mgr.14152) 527 : cluster [DBG] pgmap v414: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T07:00:01.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:00 vm04 bash[20194]: cluster 2026-03-21T06:59:59.646801+0000 mgr.x (mgr.14152) 527 : cluster [DBG] pgmap v414: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 54 KiB/s rd, 22 KiB/s wr, 51 op/s 2026-03-21T07:00:01.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:00 vm04 bash[20194]: cluster 2026-03-21T07:00:00.000133+0000 mon.a (mon.0) 999 : cluster [INF] overall HEALTH_OK 2026-03-21T07:00:01.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:00 vm04 bash[20194]: cluster 2026-03-21T07:00:00.000133+0000 mon.a (mon.0) 999 : cluster [INF] overall HEALTH_OK 2026-03-21T07:00:03.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:02 vm02 bash[17657]: cluster 2026-03-21T07:00:01.647208+0000 mgr.x (mgr.14152) 528 : cluster [DBG] pgmap v415: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 30 KiB/s rd, 1.5 KiB/s wr, 32 op/s 2026-03-21T07:00:03.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:02 vm02 bash[17657]: cluster 2026-03-21T07:00:01.647208+0000 mgr.x (mgr.14152) 528 : cluster [DBG] pgmap v415: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 30 KiB/s rd, 1.5 KiB/s wr, 32 op/s 2026-03-21T07:00:03.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:02 vm07 bash[19945]: cluster 2026-03-21T07:00:01.647208+0000 mgr.x (mgr.14152) 528 : cluster [DBG] pgmap v415: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 30 KiB/s rd, 1.5 KiB/s wr, 32 op/s 2026-03-21T07:00:03.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:02 vm07 bash[19945]: cluster 2026-03-21T07:00:01.647208+0000 mgr.x (mgr.14152) 528 : cluster [DBG] pgmap v415: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 30 KiB/s rd, 1.5 KiB/s wr, 32 op/s 2026-03-21T07:00:03.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:02 vm04 bash[20194]: cluster 2026-03-21T07:00:01.647208+0000 mgr.x (mgr.14152) 528 : cluster [DBG] pgmap v415: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 30 KiB/s rd, 1.5 KiB/s wr, 32 op/s 2026-03-21T07:00:03.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:02 vm04 bash[20194]: cluster 2026-03-21T07:00:01.647208+0000 mgr.x (mgr.14152) 528 : cluster [DBG] pgmap v415: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 30 KiB/s rd, 1.5 KiB/s wr, 32 op/s 2026-03-21T07:00:04.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:03 vm02 bash[17657]: cluster 2026-03-21T07:00:03.647604+0000 mgr.x (mgr.14152) 529 : cluster [DBG] pgmap v416: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 12 KiB/s rd, 255 B/s wr, 14 op/s 2026-03-21T07:00:04.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:03 vm02 bash[17657]: cluster 2026-03-21T07:00:03.647604+0000 mgr.x (mgr.14152) 529 : cluster [DBG] pgmap v416: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 12 KiB/s rd, 255 B/s wr, 14 op/s 2026-03-21T07:00:04.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:03 vm07 bash[19945]: cluster 2026-03-21T07:00:03.647604+0000 mgr.x (mgr.14152) 529 : cluster [DBG] pgmap v416: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 12 KiB/s rd, 255 B/s wr, 14 op/s 2026-03-21T07:00:04.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:03 vm07 bash[19945]: cluster 2026-03-21T07:00:03.647604+0000 mgr.x (mgr.14152) 529 : cluster [DBG] pgmap v416: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 12 KiB/s rd, 255 B/s wr, 14 op/s 2026-03-21T07:00:04.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:03 vm04 bash[20194]: cluster 2026-03-21T07:00:03.647604+0000 mgr.x (mgr.14152) 529 : cluster [DBG] pgmap v416: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 12 KiB/s rd, 255 B/s wr, 14 op/s 2026-03-21T07:00:04.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:03 vm04 bash[20194]: cluster 2026-03-21T07:00:03.647604+0000 mgr.x (mgr.14152) 529 : cluster [DBG] pgmap v416: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 12 KiB/s rd, 255 B/s wr, 14 op/s 2026-03-21T07:00:07.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:06 vm04 bash[20194]: cluster 2026-03-21T07:00:05.648062+0000 mgr.x (mgr.14152) 530 : cluster [DBG] pgmap v417: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 13 KiB/s rd, 255 B/s wr, 15 op/s 2026-03-21T07:00:07.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:06 vm04 bash[20194]: cluster 2026-03-21T07:00:05.648062+0000 mgr.x (mgr.14152) 530 : cluster [DBG] pgmap v417: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 13 KiB/s rd, 255 B/s wr, 15 op/s 2026-03-21T07:00:07.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:00:06 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:00:07.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:06 vm02 bash[17657]: cluster 2026-03-21T07:00:05.648062+0000 mgr.x (mgr.14152) 530 : cluster [DBG] pgmap v417: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 13 KiB/s rd, 255 B/s wr, 15 op/s 2026-03-21T07:00:07.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:06 vm02 bash[17657]: cluster 2026-03-21T07:00:05.648062+0000 mgr.x (mgr.14152) 530 : cluster [DBG] pgmap v417: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 13 KiB/s rd, 255 B/s wr, 15 op/s 2026-03-21T07:00:07.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:06 vm07 bash[19945]: cluster 2026-03-21T07:00:05.648062+0000 mgr.x (mgr.14152) 530 : cluster [DBG] pgmap v417: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 13 KiB/s rd, 255 B/s wr, 15 op/s 2026-03-21T07:00:07.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:06 vm07 bash[19945]: cluster 2026-03-21T07:00:05.648062+0000 mgr.x (mgr.14152) 530 : cluster [DBG] pgmap v417: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 13 KiB/s rd, 255 B/s wr, 15 op/s 2026-03-21T07:00:08.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:00:07 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:00:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:08 vm04 bash[20194]: audit 2026-03-21T07:00:06.826028+0000 mgr.x (mgr.14152) 531 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:08 vm04 bash[20194]: audit 2026-03-21T07:00:06.826028+0000 mgr.x (mgr.14152) 531 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:08 vm04 bash[20194]: cluster 2026-03-21T07:00:07.648428+0000 mgr.x (mgr.14152) 532 : cluster [DBG] pgmap v418: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:08 vm04 bash[20194]: cluster 2026-03-21T07:00:07.648428+0000 mgr.x (mgr.14152) 532 : cluster [DBG] pgmap v418: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:08 vm04 bash[20194]: audit 2026-03-21T07:00:07.666443+0000 mgr.x (mgr.14152) 533 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:09.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:08 vm04 bash[20194]: audit 2026-03-21T07:00:07.666443+0000 mgr.x (mgr.14152) 533 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:09.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:08 vm02 bash[17657]: audit 2026-03-21T07:00:06.826028+0000 mgr.x (mgr.14152) 531 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:09.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:08 vm02 bash[17657]: audit 2026-03-21T07:00:06.826028+0000 mgr.x (mgr.14152) 531 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:09.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:08 vm02 bash[17657]: cluster 2026-03-21T07:00:07.648428+0000 mgr.x (mgr.14152) 532 : cluster [DBG] pgmap v418: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:09.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:08 vm02 bash[17657]: cluster 2026-03-21T07:00:07.648428+0000 mgr.x (mgr.14152) 532 : cluster [DBG] pgmap v418: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:09.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:08 vm02 bash[17657]: audit 2026-03-21T07:00:07.666443+0000 mgr.x (mgr.14152) 533 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:09.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:08 vm02 bash[17657]: audit 2026-03-21T07:00:07.666443+0000 mgr.x (mgr.14152) 533 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:08 vm07 bash[19945]: audit 2026-03-21T07:00:06.826028+0000 mgr.x (mgr.14152) 531 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:08 vm07 bash[19945]: audit 2026-03-21T07:00:06.826028+0000 mgr.x (mgr.14152) 531 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:08 vm07 bash[19945]: cluster 2026-03-21T07:00:07.648428+0000 mgr.x (mgr.14152) 532 : cluster [DBG] pgmap v418: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:08 vm07 bash[19945]: cluster 2026-03-21T07:00:07.648428+0000 mgr.x (mgr.14152) 532 : cluster [DBG] pgmap v418: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:08 vm07 bash[19945]: audit 2026-03-21T07:00:07.666443+0000 mgr.x (mgr.14152) 533 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:09.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:08 vm07 bash[19945]: audit 2026-03-21T07:00:07.666443+0000 mgr.x (mgr.14152) 533 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:11.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:10 vm04 bash[20194]: cluster 2026-03-21T07:00:09.648726+0000 mgr.x (mgr.14152) 534 : cluster [DBG] pgmap v419: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:11.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:10 vm04 bash[20194]: cluster 2026-03-21T07:00:09.648726+0000 mgr.x (mgr.14152) 534 : cluster [DBG] pgmap v419: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:11.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:10 vm02 bash[17657]: cluster 2026-03-21T07:00:09.648726+0000 mgr.x (mgr.14152) 534 : cluster [DBG] pgmap v419: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:11.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:10 vm02 bash[17657]: cluster 2026-03-21T07:00:09.648726+0000 mgr.x (mgr.14152) 534 : cluster [DBG] pgmap v419: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:10 vm07 bash[19945]: cluster 2026-03-21T07:00:09.648726+0000 mgr.x (mgr.14152) 534 : cluster [DBG] pgmap v419: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:10 vm07 bash[19945]: cluster 2026-03-21T07:00:09.648726+0000 mgr.x (mgr.14152) 534 : cluster [DBG] pgmap v419: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:13.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:12 vm02 bash[17657]: cluster 2026-03-21T07:00:11.649063+0000 mgr.x (mgr.14152) 535 : cluster [DBG] pgmap v420: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:13.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:12 vm02 bash[17657]: cluster 2026-03-21T07:00:11.649063+0000 mgr.x (mgr.14152) 535 : cluster [DBG] pgmap v420: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:13.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:12 vm07 bash[19945]: cluster 2026-03-21T07:00:11.649063+0000 mgr.x (mgr.14152) 535 : cluster [DBG] pgmap v420: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:13.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:12 vm07 bash[19945]: cluster 2026-03-21T07:00:11.649063+0000 mgr.x (mgr.14152) 535 : cluster [DBG] pgmap v420: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:13.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:12 vm04 bash[20194]: cluster 2026-03-21T07:00:11.649063+0000 mgr.x (mgr.14152) 535 : cluster [DBG] pgmap v420: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:13.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:12 vm04 bash[20194]: cluster 2026-03-21T07:00:11.649063+0000 mgr.x (mgr.14152) 535 : cluster [DBG] pgmap v420: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:15.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:14 vm02 bash[17657]: cluster 2026-03-21T07:00:13.649357+0000 mgr.x (mgr.14152) 536 : cluster [DBG] pgmap v421: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:15.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:14 vm02 bash[17657]: cluster 2026-03-21T07:00:13.649357+0000 mgr.x (mgr.14152) 536 : cluster [DBG] pgmap v421: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:15.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:14 vm07 bash[19945]: cluster 2026-03-21T07:00:13.649357+0000 mgr.x (mgr.14152) 536 : cluster [DBG] pgmap v421: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:15.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:14 vm07 bash[19945]: cluster 2026-03-21T07:00:13.649357+0000 mgr.x (mgr.14152) 536 : cluster [DBG] pgmap v421: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:15.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:14 vm04 bash[20194]: cluster 2026-03-21T07:00:13.649357+0000 mgr.x (mgr.14152) 536 : cluster [DBG] pgmap v421: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:15.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:14 vm04 bash[20194]: cluster 2026-03-21T07:00:13.649357+0000 mgr.x (mgr.14152) 536 : cluster [DBG] pgmap v421: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:17.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:00:16 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:00:17.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:16 vm02 bash[17657]: cluster 2026-03-21T07:00:15.649659+0000 mgr.x (mgr.14152) 537 : cluster [DBG] pgmap v422: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:17.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:16 vm02 bash[17657]: cluster 2026-03-21T07:00:15.649659+0000 mgr.x (mgr.14152) 537 : cluster [DBG] pgmap v422: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:17.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:16 vm04 bash[20194]: cluster 2026-03-21T07:00:15.649659+0000 mgr.x (mgr.14152) 537 : cluster [DBG] pgmap v422: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:17.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:16 vm04 bash[20194]: cluster 2026-03-21T07:00:15.649659+0000 mgr.x (mgr.14152) 537 : cluster [DBG] pgmap v422: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:17.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:16 vm07 bash[19945]: cluster 2026-03-21T07:00:15.649659+0000 mgr.x (mgr.14152) 537 : cluster [DBG] pgmap v422: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:17.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:16 vm07 bash[19945]: cluster 2026-03-21T07:00:15.649659+0000 mgr.x (mgr.14152) 537 : cluster [DBG] pgmap v422: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:17.967 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:00:17 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:00:18.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:17 vm04 bash[20194]: audit 2026-03-21T07:00:16.835467+0000 mgr.x (mgr.14152) 538 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:18.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:17 vm04 bash[20194]: audit 2026-03-21T07:00:16.835467+0000 mgr.x (mgr.14152) 538 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:18.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:17 vm04 bash[20194]: cluster 2026-03-21T07:00:17.650082+0000 mgr.x (mgr.14152) 539 : cluster [DBG] pgmap v423: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:18.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:17 vm04 bash[20194]: cluster 2026-03-21T07:00:17.650082+0000 mgr.x (mgr.14152) 539 : cluster [DBG] pgmap v423: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:18.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:17 vm04 bash[20194]: audit 2026-03-21T07:00:17.677505+0000 mgr.x (mgr.14152) 540 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:18.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:17 vm04 bash[20194]: audit 2026-03-21T07:00:17.677505+0000 mgr.x (mgr.14152) 540 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:18.386 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:17 vm02 bash[17657]: audit 2026-03-21T07:00:16.835467+0000 mgr.x (mgr.14152) 538 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:18.386 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:17 vm02 bash[17657]: audit 2026-03-21T07:00:16.835467+0000 mgr.x (mgr.14152) 538 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:18.386 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:17 vm02 bash[17657]: cluster 2026-03-21T07:00:17.650082+0000 mgr.x (mgr.14152) 539 : cluster [DBG] pgmap v423: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:18.386 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:17 vm02 bash[17657]: cluster 2026-03-21T07:00:17.650082+0000 mgr.x (mgr.14152) 539 : cluster [DBG] pgmap v423: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:18.386 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:17 vm02 bash[17657]: audit 2026-03-21T07:00:17.677505+0000 mgr.x (mgr.14152) 540 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:18.386 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:17 vm02 bash[17657]: audit 2026-03-21T07:00:17.677505+0000 mgr.x (mgr.14152) 540 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:18.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:17 vm07 bash[19945]: audit 2026-03-21T07:00:16.835467+0000 mgr.x (mgr.14152) 538 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:18.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:17 vm07 bash[19945]: audit 2026-03-21T07:00:16.835467+0000 mgr.x (mgr.14152) 538 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:18.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:17 vm07 bash[19945]: cluster 2026-03-21T07:00:17.650082+0000 mgr.x (mgr.14152) 539 : cluster [DBG] pgmap v423: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:18.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:17 vm07 bash[19945]: cluster 2026-03-21T07:00:17.650082+0000 mgr.x (mgr.14152) 539 : cluster [DBG] pgmap v423: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:18.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:17 vm07 bash[19945]: audit 2026-03-21T07:00:17.677505+0000 mgr.x (mgr.14152) 540 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:18.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:17 vm07 bash[19945]: audit 2026-03-21T07:00:17.677505+0000 mgr.x (mgr.14152) 540 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:20 vm04 bash[20194]: cluster 2026-03-21T07:00:19.650505+0000 mgr.x (mgr.14152) 541 : cluster [DBG] pgmap v424: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:20 vm04 bash[20194]: cluster 2026-03-21T07:00:19.650505+0000 mgr.x (mgr.14152) 541 : cluster [DBG] pgmap v424: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:21.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:20 vm02 bash[17657]: cluster 2026-03-21T07:00:19.650505+0000 mgr.x (mgr.14152) 541 : cluster [DBG] pgmap v424: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:21.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:20 vm02 bash[17657]: cluster 2026-03-21T07:00:19.650505+0000 mgr.x (mgr.14152) 541 : cluster [DBG] pgmap v424: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:20 vm07 bash[19945]: cluster 2026-03-21T07:00:19.650505+0000 mgr.x (mgr.14152) 541 : cluster [DBG] pgmap v424: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:20 vm07 bash[19945]: cluster 2026-03-21T07:00:19.650505+0000 mgr.x (mgr.14152) 541 : cluster [DBG] pgmap v424: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:23.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:22 vm04 bash[20194]: cluster 2026-03-21T07:00:21.650944+0000 mgr.x (mgr.14152) 542 : cluster [DBG] pgmap v425: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:23.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:22 vm04 bash[20194]: cluster 2026-03-21T07:00:21.650944+0000 mgr.x (mgr.14152) 542 : cluster [DBG] pgmap v425: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:23.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:22 vm02 bash[17657]: cluster 2026-03-21T07:00:21.650944+0000 mgr.x (mgr.14152) 542 : cluster [DBG] pgmap v425: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:23.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:22 vm02 bash[17657]: cluster 2026-03-21T07:00:21.650944+0000 mgr.x (mgr.14152) 542 : cluster [DBG] pgmap v425: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:22 vm07 bash[19945]: cluster 2026-03-21T07:00:21.650944+0000 mgr.x (mgr.14152) 542 : cluster [DBG] pgmap v425: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:22 vm07 bash[19945]: cluster 2026-03-21T07:00:21.650944+0000 mgr.x (mgr.14152) 542 : cluster [DBG] pgmap v425: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:25.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:24 vm04 bash[20194]: cluster 2026-03-21T07:00:23.651359+0000 mgr.x (mgr.14152) 543 : cluster [DBG] pgmap v426: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:25.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:24 vm04 bash[20194]: cluster 2026-03-21T07:00:23.651359+0000 mgr.x (mgr.14152) 543 : cluster [DBG] pgmap v426: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:25.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:24 vm02 bash[17657]: cluster 2026-03-21T07:00:23.651359+0000 mgr.x (mgr.14152) 543 : cluster [DBG] pgmap v426: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:25.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:24 vm02 bash[17657]: cluster 2026-03-21T07:00:23.651359+0000 mgr.x (mgr.14152) 543 : cluster [DBG] pgmap v426: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:24 vm07 bash[19945]: cluster 2026-03-21T07:00:23.651359+0000 mgr.x (mgr.14152) 543 : cluster [DBG] pgmap v426: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:24 vm07 bash[19945]: cluster 2026-03-21T07:00:23.651359+0000 mgr.x (mgr.14152) 543 : cluster [DBG] pgmap v426: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:27.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:26 vm04 bash[20194]: cluster 2026-03-21T07:00:25.651760+0000 mgr.x (mgr.14152) 544 : cluster [DBG] pgmap v427: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:27.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:26 vm04 bash[20194]: cluster 2026-03-21T07:00:25.651760+0000 mgr.x (mgr.14152) 544 : cluster [DBG] pgmap v427: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:27.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:00:26 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:00:27.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:26 vm02 bash[17657]: cluster 2026-03-21T07:00:25.651760+0000 mgr.x (mgr.14152) 544 : cluster [DBG] pgmap v427: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:27.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:26 vm02 bash[17657]: cluster 2026-03-21T07:00:25.651760+0000 mgr.x (mgr.14152) 544 : cluster [DBG] pgmap v427: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:26 vm07 bash[19945]: cluster 2026-03-21T07:00:25.651760+0000 mgr.x (mgr.14152) 544 : cluster [DBG] pgmap v427: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:26 vm07 bash[19945]: cluster 2026-03-21T07:00:25.651760+0000 mgr.x (mgr.14152) 544 : cluster [DBG] pgmap v427: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:28.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:27 vm04 bash[20194]: audit 2026-03-21T07:00:27.689962+0000 mon.a (mon.0) 1000 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:00:28.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:27 vm04 bash[20194]: audit 2026-03-21T07:00:27.689962+0000 mon.a (mon.0) 1000 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:00:28.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:27 vm02 bash[17657]: audit 2026-03-21T07:00:27.689962+0000 mon.a (mon.0) 1000 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:00:28.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:27 vm02 bash[17657]: audit 2026-03-21T07:00:27.689962+0000 mon.a (mon.0) 1000 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:00:28.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:00:27 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:00:28.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:27 vm07 bash[19945]: audit 2026-03-21T07:00:27.689962+0000 mon.a (mon.0) 1000 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:00:28.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:27 vm07 bash[19945]: audit 2026-03-21T07:00:27.689962+0000 mon.a (mon.0) 1000 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:26.843531+0000 mgr.x (mgr.14152) 545 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:26.843531+0000 mgr.x (mgr.14152) 545 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: cluster 2026-03-21T07:00:27.652075+0000 mgr.x (mgr.14152) 546 : cluster [DBG] pgmap v428: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: cluster 2026-03-21T07:00:27.652075+0000 mgr.x (mgr.14152) 546 : cluster [DBG] pgmap v428: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:27.688310+0000 mgr.x (mgr.14152) 547 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:27.688310+0000 mgr.x (mgr.14152) 547 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:28.062303+0000 mon.a (mon.0) 1001 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:28.062303+0000 mon.a (mon.0) 1001 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:28.062909+0000 mon.a (mon.0) 1002 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:28.062909+0000 mon.a (mon.0) 1002 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:28.068058+0000 mon.a (mon.0) 1003 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:00:29.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:28.068058+0000 mon.a (mon.0) 1003 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:00:29.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:28.068879+0000 mon.a (mon.0) 1004 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:00:29.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:28 vm04 bash[20194]: audit 2026-03-21T07:00:28.068879+0000 mon.a (mon.0) 1004 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:00:29.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:26.843531+0000 mgr.x (mgr.14152) 545 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:26.843531+0000 mgr.x (mgr.14152) 545 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: cluster 2026-03-21T07:00:27.652075+0000 mgr.x (mgr.14152) 546 : cluster [DBG] pgmap v428: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:29.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: cluster 2026-03-21T07:00:27.652075+0000 mgr.x (mgr.14152) 546 : cluster [DBG] pgmap v428: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:29.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:27.688310+0000 mgr.x (mgr.14152) 547 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:27.688310+0000 mgr.x (mgr.14152) 547 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:28.062303+0000 mon.a (mon.0) 1001 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:00:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:28.062303+0000 mon.a (mon.0) 1001 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:00:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:28.062909+0000 mon.a (mon.0) 1002 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:00:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:28.062909+0000 mon.a (mon.0) 1002 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:00:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:28.068058+0000 mon.a (mon.0) 1003 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:00:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:28.068058+0000 mon.a (mon.0) 1003 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:00:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:28.068879+0000 mon.a (mon.0) 1004 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:00:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:28 vm02 bash[17657]: audit 2026-03-21T07:00:28.068879+0000 mon.a (mon.0) 1004 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:26.843531+0000 mgr.x (mgr.14152) 545 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:26.843531+0000 mgr.x (mgr.14152) 545 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: cluster 2026-03-21T07:00:27.652075+0000 mgr.x (mgr.14152) 546 : cluster [DBG] pgmap v428: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: cluster 2026-03-21T07:00:27.652075+0000 mgr.x (mgr.14152) 546 : cluster [DBG] pgmap v428: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:27.688310+0000 mgr.x (mgr.14152) 547 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:27.688310+0000 mgr.x (mgr.14152) 547 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:28.062303+0000 mon.a (mon.0) 1001 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:28.062303+0000 mon.a (mon.0) 1001 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:28.062909+0000 mon.a (mon.0) 1002 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:28.062909+0000 mon.a (mon.0) 1002 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:28.068058+0000 mon.a (mon.0) 1003 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:28.068058+0000 mon.a (mon.0) 1003 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:28.068879+0000 mon.a (mon.0) 1004 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:00:29.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:28 vm07 bash[19945]: audit 2026-03-21T07:00:28.068879+0000 mon.a (mon.0) 1004 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:00:31.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:30 vm02 bash[17657]: cluster 2026-03-21T07:00:29.652609+0000 mgr.x (mgr.14152) 548 : cluster [DBG] pgmap v429: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:31.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:30 vm02 bash[17657]: cluster 2026-03-21T07:00:29.652609+0000 mgr.x (mgr.14152) 548 : cluster [DBG] pgmap v429: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:30 vm07 bash[19945]: cluster 2026-03-21T07:00:29.652609+0000 mgr.x (mgr.14152) 548 : cluster [DBG] pgmap v429: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:30 vm07 bash[19945]: cluster 2026-03-21T07:00:29.652609+0000 mgr.x (mgr.14152) 548 : cluster [DBG] pgmap v429: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:31.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:30 vm04 bash[20194]: cluster 2026-03-21T07:00:29.652609+0000 mgr.x (mgr.14152) 548 : cluster [DBG] pgmap v429: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:31.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:30 vm04 bash[20194]: cluster 2026-03-21T07:00:29.652609+0000 mgr.x (mgr.14152) 548 : cluster [DBG] pgmap v429: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:31.666 INFO:teuthology.orchestra.run.vm02.stderr:Note: switching to '0392f78529848ec72469e8e431875cb98d3a5fb4'. 2026-03-21T07:00:31.666 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr:state without impacting any branches by switching back to a branch. 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr: git switch -c 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr:Or undo this operation with: 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr: git switch - 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T07:00:31.667 INFO:teuthology.orchestra.run.vm02.stderr:HEAD is now at 0392f785298 qa/tasks/keystone: restart mariadb for rocky and alma linux too 2026-03-21T07:00:31.675 DEBUG:teuthology.orchestra.run.vm02:> cp -- /home/ubuntu/cephtest/clone.client.0/src/test/cli-integration/rbd/rest_api_create.t /home/ubuntu/cephtest/archive/cram.client.0 2026-03-21T07:00:31.722 DEBUG:teuthology.orchestra.run.vm04:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.1 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-21T07:00:32.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:31 vm02 bash[17657]: cluster 2026-03-21T07:00:31.653052+0000 mgr.x (mgr.14152) 549 : cluster [DBG] pgmap v430: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:32.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:31 vm02 bash[17657]: cluster 2026-03-21T07:00:31.653052+0000 mgr.x (mgr.14152) 549 : cluster [DBG] pgmap v430: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:31 vm07 bash[19945]: cluster 2026-03-21T07:00:31.653052+0000 mgr.x (mgr.14152) 549 : cluster [DBG] pgmap v430: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:31 vm07 bash[19945]: cluster 2026-03-21T07:00:31.653052+0000 mgr.x (mgr.14152) 549 : cluster [DBG] pgmap v430: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:31 vm04 bash[20194]: cluster 2026-03-21T07:00:31.653052+0000 mgr.x (mgr.14152) 549 : cluster [DBG] pgmap v430: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:31 vm04 bash[20194]: cluster 2026-03-21T07:00:31.653052+0000 mgr.x (mgr.14152) 549 : cluster [DBG] pgmap v430: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:33.398 INFO:teuthology.orchestra.run.vm04.stdout:Collecting cram==0.6 2026-03-21T07:00:33.399 INFO:teuthology.orchestra.run.vm04.stdout: Using cached cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-21T07:00:33.410 INFO:teuthology.orchestra.run.vm04.stdout:Installing collected packages: cram 2026-03-21T07:00:33.415 INFO:teuthology.orchestra.run.vm04.stdout:Successfully installed cram-0.6 2026-03-21T07:00:33.445 DEBUG:teuthology.orchestra.run.vm04:> rm -rf /home/ubuntu/cephtest/clone.client.1 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.1 && cd /home/ubuntu/cephtest/clone.client.1 && git checkout 0392f78529848ec72469e8e431875cb98d3a5fb4 2026-03-21T07:00:33.449 INFO:teuthology.orchestra.run.vm04.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.1'... 2026-03-21T07:00:35.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:34 vm04 bash[20194]: cluster 2026-03-21T07:00:33.653376+0000 mgr.x (mgr.14152) 550 : cluster [DBG] pgmap v431: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:35.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:34 vm04 bash[20194]: cluster 2026-03-21T07:00:33.653376+0000 mgr.x (mgr.14152) 550 : cluster [DBG] pgmap v431: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:35.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:34 vm02 bash[17657]: cluster 2026-03-21T07:00:33.653376+0000 mgr.x (mgr.14152) 550 : cluster [DBG] pgmap v431: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:35.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:34 vm02 bash[17657]: cluster 2026-03-21T07:00:33.653376+0000 mgr.x (mgr.14152) 550 : cluster [DBG] pgmap v431: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:35.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:34 vm07 bash[19945]: cluster 2026-03-21T07:00:33.653376+0000 mgr.x (mgr.14152) 550 : cluster [DBG] pgmap v431: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:35.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:34 vm07 bash[19945]: cluster 2026-03-21T07:00:33.653376+0000 mgr.x (mgr.14152) 550 : cluster [DBG] pgmap v431: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:37.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:36 vm04 bash[20194]: cluster 2026-03-21T07:00:35.653729+0000 mgr.x (mgr.14152) 551 : cluster [DBG] pgmap v432: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:37.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:36 vm04 bash[20194]: cluster 2026-03-21T07:00:35.653729+0000 mgr.x (mgr.14152) 551 : cluster [DBG] pgmap v432: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:37.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:36 vm02 bash[17657]: cluster 2026-03-21T07:00:35.653729+0000 mgr.x (mgr.14152) 551 : cluster [DBG] pgmap v432: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:37.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:36 vm02 bash[17657]: cluster 2026-03-21T07:00:35.653729+0000 mgr.x (mgr.14152) 551 : cluster [DBG] pgmap v432: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:37.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:00:36 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:00:37.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:36 vm07 bash[19945]: cluster 2026-03-21T07:00:35.653729+0000 mgr.x (mgr.14152) 551 : cluster [DBG] pgmap v432: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:37.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:36 vm07 bash[19945]: cluster 2026-03-21T07:00:35.653729+0000 mgr.x (mgr.14152) 551 : cluster [DBG] pgmap v432: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:38.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:00:37 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:00:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:38 vm04 bash[20194]: audit 2026-03-21T07:00:36.854278+0000 mgr.x (mgr.14152) 552 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:38 vm04 bash[20194]: audit 2026-03-21T07:00:36.854278+0000 mgr.x (mgr.14152) 552 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:38 vm04 bash[20194]: cluster 2026-03-21T07:00:37.654093+0000 mgr.x (mgr.14152) 553 : cluster [DBG] pgmap v433: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:38 vm04 bash[20194]: cluster 2026-03-21T07:00:37.654093+0000 mgr.x (mgr.14152) 553 : cluster [DBG] pgmap v433: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:38 vm04 bash[20194]: audit 2026-03-21T07:00:37.694766+0000 mgr.x (mgr.14152) 554 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:39.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:38 vm04 bash[20194]: audit 2026-03-21T07:00:37.694766+0000 mgr.x (mgr.14152) 554 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:39.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:38 vm02 bash[17657]: audit 2026-03-21T07:00:36.854278+0000 mgr.x (mgr.14152) 552 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:39.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:38 vm02 bash[17657]: audit 2026-03-21T07:00:36.854278+0000 mgr.x (mgr.14152) 552 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:39.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:38 vm02 bash[17657]: cluster 2026-03-21T07:00:37.654093+0000 mgr.x (mgr.14152) 553 : cluster [DBG] pgmap v433: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:39.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:38 vm02 bash[17657]: cluster 2026-03-21T07:00:37.654093+0000 mgr.x (mgr.14152) 553 : cluster [DBG] pgmap v433: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:39.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:38 vm02 bash[17657]: audit 2026-03-21T07:00:37.694766+0000 mgr.x (mgr.14152) 554 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:39.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:38 vm02 bash[17657]: audit 2026-03-21T07:00:37.694766+0000 mgr.x (mgr.14152) 554 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:38 vm07 bash[19945]: audit 2026-03-21T07:00:36.854278+0000 mgr.x (mgr.14152) 552 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:39.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:38 vm07 bash[19945]: audit 2026-03-21T07:00:36.854278+0000 mgr.x (mgr.14152) 552 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:39.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:38 vm07 bash[19945]: cluster 2026-03-21T07:00:37.654093+0000 mgr.x (mgr.14152) 553 : cluster [DBG] pgmap v433: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:39.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:38 vm07 bash[19945]: cluster 2026-03-21T07:00:37.654093+0000 mgr.x (mgr.14152) 553 : cluster [DBG] pgmap v433: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:39.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:38 vm07 bash[19945]: audit 2026-03-21T07:00:37.694766+0000 mgr.x (mgr.14152) 554 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:39.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:38 vm07 bash[19945]: audit 2026-03-21T07:00:37.694766+0000 mgr.x (mgr.14152) 554 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:41.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:40 vm04 bash[20194]: cluster 2026-03-21T07:00:39.654498+0000 mgr.x (mgr.14152) 555 : cluster [DBG] pgmap v434: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:41.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:40 vm04 bash[20194]: cluster 2026-03-21T07:00:39.654498+0000 mgr.x (mgr.14152) 555 : cluster [DBG] pgmap v434: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:41.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:40 vm02 bash[17657]: cluster 2026-03-21T07:00:39.654498+0000 mgr.x (mgr.14152) 555 : cluster [DBG] pgmap v434: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:41.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:40 vm02 bash[17657]: cluster 2026-03-21T07:00:39.654498+0000 mgr.x (mgr.14152) 555 : cluster [DBG] pgmap v434: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:41.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:40 vm07 bash[19945]: cluster 2026-03-21T07:00:39.654498+0000 mgr.x (mgr.14152) 555 : cluster [DBG] pgmap v434: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:41.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:40 vm07 bash[19945]: cluster 2026-03-21T07:00:39.654498+0000 mgr.x (mgr.14152) 555 : cluster [DBG] pgmap v434: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:43.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:42 vm04 bash[20194]: cluster 2026-03-21T07:00:41.654873+0000 mgr.x (mgr.14152) 556 : cluster [DBG] pgmap v435: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:43.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:42 vm04 bash[20194]: cluster 2026-03-21T07:00:41.654873+0000 mgr.x (mgr.14152) 556 : cluster [DBG] pgmap v435: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:43.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:42 vm02 bash[17657]: cluster 2026-03-21T07:00:41.654873+0000 mgr.x (mgr.14152) 556 : cluster [DBG] pgmap v435: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:43.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:42 vm02 bash[17657]: cluster 2026-03-21T07:00:41.654873+0000 mgr.x (mgr.14152) 556 : cluster [DBG] pgmap v435: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:43.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:42 vm07 bash[19945]: cluster 2026-03-21T07:00:41.654873+0000 mgr.x (mgr.14152) 556 : cluster [DBG] pgmap v435: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:43.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:42 vm07 bash[19945]: cluster 2026-03-21T07:00:41.654873+0000 mgr.x (mgr.14152) 556 : cluster [DBG] pgmap v435: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:44.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:43 vm02 bash[17657]: cluster 2026-03-21T07:00:43.655331+0000 mgr.x (mgr.14152) 557 : cluster [DBG] pgmap v436: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:44.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:43 vm02 bash[17657]: cluster 2026-03-21T07:00:43.655331+0000 mgr.x (mgr.14152) 557 : cluster [DBG] pgmap v436: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:44.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:43 vm07 bash[19945]: cluster 2026-03-21T07:00:43.655331+0000 mgr.x (mgr.14152) 557 : cluster [DBG] pgmap v436: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:44.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:43 vm07 bash[19945]: cluster 2026-03-21T07:00:43.655331+0000 mgr.x (mgr.14152) 557 : cluster [DBG] pgmap v436: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:44.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:43 vm04 bash[20194]: cluster 2026-03-21T07:00:43.655331+0000 mgr.x (mgr.14152) 557 : cluster [DBG] pgmap v436: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:44.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:43 vm04 bash[20194]: cluster 2026-03-21T07:00:43.655331+0000 mgr.x (mgr.14152) 557 : cluster [DBG] pgmap v436: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:47.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:46 vm04 bash[20194]: cluster 2026-03-21T07:00:45.655745+0000 mgr.x (mgr.14152) 558 : cluster [DBG] pgmap v437: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:47.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:46 vm04 bash[20194]: cluster 2026-03-21T07:00:45.655745+0000 mgr.x (mgr.14152) 558 : cluster [DBG] pgmap v437: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:47.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:46 vm02 bash[17657]: cluster 2026-03-21T07:00:45.655745+0000 mgr.x (mgr.14152) 558 : cluster [DBG] pgmap v437: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:47.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:46 vm02 bash[17657]: cluster 2026-03-21T07:00:45.655745+0000 mgr.x (mgr.14152) 558 : cluster [DBG] pgmap v437: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:47.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:00:46 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:00:47.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:46 vm07 bash[19945]: cluster 2026-03-21T07:00:45.655745+0000 mgr.x (mgr.14152) 558 : cluster [DBG] pgmap v437: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:47.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:46 vm07 bash[19945]: cluster 2026-03-21T07:00:45.655745+0000 mgr.x (mgr.14152) 558 : cluster [DBG] pgmap v437: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:48.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:00:47 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:00:49.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:49 vm02 bash[17657]: audit 2026-03-21T07:00:46.864825+0000 mgr.x (mgr.14152) 559 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:49.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:49 vm02 bash[17657]: audit 2026-03-21T07:00:46.864825+0000 mgr.x (mgr.14152) 559 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:49.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:49 vm02 bash[17657]: cluster 2026-03-21T07:00:47.656054+0000 mgr.x (mgr.14152) 560 : cluster [DBG] pgmap v438: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:49.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:49 vm02 bash[17657]: cluster 2026-03-21T07:00:47.656054+0000 mgr.x (mgr.14152) 560 : cluster [DBG] pgmap v438: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:49.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:49 vm02 bash[17657]: audit 2026-03-21T07:00:47.705536+0000 mgr.x (mgr.14152) 561 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:49.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:49 vm02 bash[17657]: audit 2026-03-21T07:00:47.705536+0000 mgr.x (mgr.14152) 561 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:49.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:49 vm07 bash[19945]: audit 2026-03-21T07:00:46.864825+0000 mgr.x (mgr.14152) 559 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:49.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:49 vm07 bash[19945]: audit 2026-03-21T07:00:46.864825+0000 mgr.x (mgr.14152) 559 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:49.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:49 vm07 bash[19945]: cluster 2026-03-21T07:00:47.656054+0000 mgr.x (mgr.14152) 560 : cluster [DBG] pgmap v438: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:49.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:49 vm07 bash[19945]: cluster 2026-03-21T07:00:47.656054+0000 mgr.x (mgr.14152) 560 : cluster [DBG] pgmap v438: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:49.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:49 vm07 bash[19945]: audit 2026-03-21T07:00:47.705536+0000 mgr.x (mgr.14152) 561 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:49.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:49 vm07 bash[19945]: audit 2026-03-21T07:00:47.705536+0000 mgr.x (mgr.14152) 561 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:49.509 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:49 vm04 bash[20194]: audit 2026-03-21T07:00:46.864825+0000 mgr.x (mgr.14152) 559 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:49.509 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:49 vm04 bash[20194]: audit 2026-03-21T07:00:46.864825+0000 mgr.x (mgr.14152) 559 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:49.509 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:49 vm04 bash[20194]: cluster 2026-03-21T07:00:47.656054+0000 mgr.x (mgr.14152) 560 : cluster [DBG] pgmap v438: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:49.509 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:49 vm04 bash[20194]: cluster 2026-03-21T07:00:47.656054+0000 mgr.x (mgr.14152) 560 : cluster [DBG] pgmap v438: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:49.509 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:49 vm04 bash[20194]: audit 2026-03-21T07:00:47.705536+0000 mgr.x (mgr.14152) 561 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:49.509 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:49 vm04 bash[20194]: audit 2026-03-21T07:00:47.705536+0000 mgr.x (mgr.14152) 561 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:50.386 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:50 vm02 bash[17657]: cluster 2026-03-21T07:00:49.656381+0000 mgr.x (mgr.14152) 562 : cluster [DBG] pgmap v439: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:50.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:50 vm02 bash[17657]: cluster 2026-03-21T07:00:49.656381+0000 mgr.x (mgr.14152) 562 : cluster [DBG] pgmap v439: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:50.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:50 vm07 bash[19945]: cluster 2026-03-21T07:00:49.656381+0000 mgr.x (mgr.14152) 562 : cluster [DBG] pgmap v439: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:50.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:50 vm07 bash[19945]: cluster 2026-03-21T07:00:49.656381+0000 mgr.x (mgr.14152) 562 : cluster [DBG] pgmap v439: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:50.509 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:50 vm04 bash[20194]: cluster 2026-03-21T07:00:49.656381+0000 mgr.x (mgr.14152) 562 : cluster [DBG] pgmap v439: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:50.509 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:50 vm04 bash[20194]: cluster 2026-03-21T07:00:49.656381+0000 mgr.x (mgr.14152) 562 : cluster [DBG] pgmap v439: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:53.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:52 vm04 bash[20194]: cluster 2026-03-21T07:00:51.656678+0000 mgr.x (mgr.14152) 563 : cluster [DBG] pgmap v440: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:53.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:52 vm04 bash[20194]: cluster 2026-03-21T07:00:51.656678+0000 mgr.x (mgr.14152) 563 : cluster [DBG] pgmap v440: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:53.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:52 vm02 bash[17657]: cluster 2026-03-21T07:00:51.656678+0000 mgr.x (mgr.14152) 563 : cluster [DBG] pgmap v440: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:53.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:52 vm02 bash[17657]: cluster 2026-03-21T07:00:51.656678+0000 mgr.x (mgr.14152) 563 : cluster [DBG] pgmap v440: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:52 vm07 bash[19945]: cluster 2026-03-21T07:00:51.656678+0000 mgr.x (mgr.14152) 563 : cluster [DBG] pgmap v440: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:53.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:52 vm07 bash[19945]: cluster 2026-03-21T07:00:51.656678+0000 mgr.x (mgr.14152) 563 : cluster [DBG] pgmap v440: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:55.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:54 vm04 bash[20194]: cluster 2026-03-21T07:00:53.656934+0000 mgr.x (mgr.14152) 564 : cluster [DBG] pgmap v441: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:55.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:54 vm04 bash[20194]: cluster 2026-03-21T07:00:53.656934+0000 mgr.x (mgr.14152) 564 : cluster [DBG] pgmap v441: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:55.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:54 vm02 bash[17657]: cluster 2026-03-21T07:00:53.656934+0000 mgr.x (mgr.14152) 564 : cluster [DBG] pgmap v441: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:55.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:54 vm02 bash[17657]: cluster 2026-03-21T07:00:53.656934+0000 mgr.x (mgr.14152) 564 : cluster [DBG] pgmap v441: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:55.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:54 vm07 bash[19945]: cluster 2026-03-21T07:00:53.656934+0000 mgr.x (mgr.14152) 564 : cluster [DBG] pgmap v441: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:55.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:54 vm07 bash[19945]: cluster 2026-03-21T07:00:53.656934+0000 mgr.x (mgr.14152) 564 : cluster [DBG] pgmap v441: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:57.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:56 vm04 bash[20194]: cluster 2026-03-21T07:00:55.657272+0000 mgr.x (mgr.14152) 565 : cluster [DBG] pgmap v442: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:57.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:56 vm04 bash[20194]: cluster 2026-03-21T07:00:55.657272+0000 mgr.x (mgr.14152) 565 : cluster [DBG] pgmap v442: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:57.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:00:56 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:00:57.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:56 vm02 bash[17657]: cluster 2026-03-21T07:00:55.657272+0000 mgr.x (mgr.14152) 565 : cluster [DBG] pgmap v442: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:57.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:56 vm02 bash[17657]: cluster 2026-03-21T07:00:55.657272+0000 mgr.x (mgr.14152) 565 : cluster [DBG] pgmap v442: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:57.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:56 vm07 bash[19945]: cluster 2026-03-21T07:00:55.657272+0000 mgr.x (mgr.14152) 565 : cluster [DBG] pgmap v442: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:57.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:56 vm07 bash[19945]: cluster 2026-03-21T07:00:55.657272+0000 mgr.x (mgr.14152) 565 : cluster [DBG] pgmap v442: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:00:58.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:00:57 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:00:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:58 vm04 bash[20194]: audit 2026-03-21T07:00:56.871458+0000 mgr.x (mgr.14152) 566 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:58 vm04 bash[20194]: audit 2026-03-21T07:00:56.871458+0000 mgr.x (mgr.14152) 566 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:58 vm04 bash[20194]: cluster 2026-03-21T07:00:57.657652+0000 mgr.x (mgr.14152) 567 : cluster [DBG] pgmap v443: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:58 vm04 bash[20194]: cluster 2026-03-21T07:00:57.657652+0000 mgr.x (mgr.14152) 567 : cluster [DBG] pgmap v443: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:58 vm04 bash[20194]: audit 2026-03-21T07:00:57.714322+0000 mgr.x (mgr.14152) 568 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:59.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:00:58 vm04 bash[20194]: audit 2026-03-21T07:00:57.714322+0000 mgr.x (mgr.14152) 568 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:58 vm02 bash[17657]: audit 2026-03-21T07:00:56.871458+0000 mgr.x (mgr.14152) 566 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:58 vm02 bash[17657]: audit 2026-03-21T07:00:56.871458+0000 mgr.x (mgr.14152) 566 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:58 vm02 bash[17657]: cluster 2026-03-21T07:00:57.657652+0000 mgr.x (mgr.14152) 567 : cluster [DBG] pgmap v443: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:58 vm02 bash[17657]: cluster 2026-03-21T07:00:57.657652+0000 mgr.x (mgr.14152) 567 : cluster [DBG] pgmap v443: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:58 vm02 bash[17657]: audit 2026-03-21T07:00:57.714322+0000 mgr.x (mgr.14152) 568 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:59.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:00:58 vm02 bash[17657]: audit 2026-03-21T07:00:57.714322+0000 mgr.x (mgr.14152) 568 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:58 vm07 bash[19945]: audit 2026-03-21T07:00:56.871458+0000 mgr.x (mgr.14152) 566 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:58 vm07 bash[19945]: audit 2026-03-21T07:00:56.871458+0000 mgr.x (mgr.14152) 566 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:58 vm07 bash[19945]: cluster 2026-03-21T07:00:57.657652+0000 mgr.x (mgr.14152) 567 : cluster [DBG] pgmap v443: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:58 vm07 bash[19945]: cluster 2026-03-21T07:00:57.657652+0000 mgr.x (mgr.14152) 567 : cluster [DBG] pgmap v443: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:00:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:58 vm07 bash[19945]: audit 2026-03-21T07:00:57.714322+0000 mgr.x (mgr.14152) 568 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:00:59.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:00:58 vm07 bash[19945]: audit 2026-03-21T07:00:57.714322+0000 mgr.x (mgr.14152) 568 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:01.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:00 vm04 bash[20194]: cluster 2026-03-21T07:00:59.657972+0000 mgr.x (mgr.14152) 569 : cluster [DBG] pgmap v444: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:01.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:00 vm04 bash[20194]: cluster 2026-03-21T07:00:59.657972+0000 mgr.x (mgr.14152) 569 : cluster [DBG] pgmap v444: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:01.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:00 vm02 bash[17657]: cluster 2026-03-21T07:00:59.657972+0000 mgr.x (mgr.14152) 569 : cluster [DBG] pgmap v444: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:01.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:00 vm02 bash[17657]: cluster 2026-03-21T07:00:59.657972+0000 mgr.x (mgr.14152) 569 : cluster [DBG] pgmap v444: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:01.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:00 vm07 bash[19945]: cluster 2026-03-21T07:00:59.657972+0000 mgr.x (mgr.14152) 569 : cluster [DBG] pgmap v444: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:01.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:00 vm07 bash[19945]: cluster 2026-03-21T07:00:59.657972+0000 mgr.x (mgr.14152) 569 : cluster [DBG] pgmap v444: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:03.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:02 vm02 bash[17657]: cluster 2026-03-21T07:01:01.658286+0000 mgr.x (mgr.14152) 570 : cluster [DBG] pgmap v445: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:03.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:02 vm02 bash[17657]: cluster 2026-03-21T07:01:01.658286+0000 mgr.x (mgr.14152) 570 : cluster [DBG] pgmap v445: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:03.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:02 vm07 bash[19945]: cluster 2026-03-21T07:01:01.658286+0000 mgr.x (mgr.14152) 570 : cluster [DBG] pgmap v445: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:03.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:02 vm07 bash[19945]: cluster 2026-03-21T07:01:01.658286+0000 mgr.x (mgr.14152) 570 : cluster [DBG] pgmap v445: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:03.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:02 vm04 bash[20194]: cluster 2026-03-21T07:01:01.658286+0000 mgr.x (mgr.14152) 570 : cluster [DBG] pgmap v445: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:03.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:02 vm04 bash[20194]: cluster 2026-03-21T07:01:01.658286+0000 mgr.x (mgr.14152) 570 : cluster [DBG] pgmap v445: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:04.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:03 vm02 bash[17657]: cluster 2026-03-21T07:01:03.658605+0000 mgr.x (mgr.14152) 571 : cluster [DBG] pgmap v446: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:04.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:03 vm02 bash[17657]: cluster 2026-03-21T07:01:03.658605+0000 mgr.x (mgr.14152) 571 : cluster [DBG] pgmap v446: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:04.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:03 vm07 bash[19945]: cluster 2026-03-21T07:01:03.658605+0000 mgr.x (mgr.14152) 571 : cluster [DBG] pgmap v446: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:04.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:03 vm07 bash[19945]: cluster 2026-03-21T07:01:03.658605+0000 mgr.x (mgr.14152) 571 : cluster [DBG] pgmap v446: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:04.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:03 vm04 bash[20194]: cluster 2026-03-21T07:01:03.658605+0000 mgr.x (mgr.14152) 571 : cluster [DBG] pgmap v446: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:04.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:03 vm04 bash[20194]: cluster 2026-03-21T07:01:03.658605+0000 mgr.x (mgr.14152) 571 : cluster [DBG] pgmap v446: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:07.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:06 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:01:07.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:06 vm02 bash[17657]: cluster 2026-03-21T07:01:05.658890+0000 mgr.x (mgr.14152) 572 : cluster [DBG] pgmap v447: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:07.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:06 vm02 bash[17657]: cluster 2026-03-21T07:01:05.658890+0000 mgr.x (mgr.14152) 572 : cluster [DBG] pgmap v447: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:07.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:06 vm07 bash[19945]: cluster 2026-03-21T07:01:05.658890+0000 mgr.x (mgr.14152) 572 : cluster [DBG] pgmap v447: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:07.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:06 vm07 bash[19945]: cluster 2026-03-21T07:01:05.658890+0000 mgr.x (mgr.14152) 572 : cluster [DBG] pgmap v447: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:07.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:06 vm04 bash[20194]: cluster 2026-03-21T07:01:05.658890+0000 mgr.x (mgr.14152) 572 : cluster [DBG] pgmap v447: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:07.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:06 vm04 bash[20194]: cluster 2026-03-21T07:01:05.658890+0000 mgr.x (mgr.14152) 572 : cluster [DBG] pgmap v447: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:08.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:07 vm02 bash[17657]: audit 2026-03-21T07:01:06.882269+0000 mgr.x (mgr.14152) 573 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:07 vm02 bash[17657]: audit 2026-03-21T07:01:06.882269+0000 mgr.x (mgr.14152) 573 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:07 vm02 bash[17657]: cluster 2026-03-21T07:01:07.659303+0000 mgr.x (mgr.14152) 574 : cluster [DBG] pgmap v448: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:08.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:07 vm02 bash[17657]: cluster 2026-03-21T07:01:07.659303+0000 mgr.x (mgr.14152) 574 : cluster [DBG] pgmap v448: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:08.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:07 vm02 bash[17657]: audit 2026-03-21T07:01:07.717746+0000 mgr.x (mgr.14152) 575 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:07 vm02 bash[17657]: audit 2026-03-21T07:01:07.717746+0000 mgr.x (mgr.14152) 575 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:07 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:01:08.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:07 vm07 bash[19945]: audit 2026-03-21T07:01:06.882269+0000 mgr.x (mgr.14152) 573 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:07 vm07 bash[19945]: audit 2026-03-21T07:01:06.882269+0000 mgr.x (mgr.14152) 573 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:07 vm07 bash[19945]: cluster 2026-03-21T07:01:07.659303+0000 mgr.x (mgr.14152) 574 : cluster [DBG] pgmap v448: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:08.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:07 vm07 bash[19945]: cluster 2026-03-21T07:01:07.659303+0000 mgr.x (mgr.14152) 574 : cluster [DBG] pgmap v448: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:08.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:07 vm07 bash[19945]: audit 2026-03-21T07:01:07.717746+0000 mgr.x (mgr.14152) 575 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:07 vm07 bash[19945]: audit 2026-03-21T07:01:07.717746+0000 mgr.x (mgr.14152) 575 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:07 vm04 bash[20194]: audit 2026-03-21T07:01:06.882269+0000 mgr.x (mgr.14152) 573 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:07 vm04 bash[20194]: audit 2026-03-21T07:01:06.882269+0000 mgr.x (mgr.14152) 573 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:07 vm04 bash[20194]: cluster 2026-03-21T07:01:07.659303+0000 mgr.x (mgr.14152) 574 : cluster [DBG] pgmap v448: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:08.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:07 vm04 bash[20194]: cluster 2026-03-21T07:01:07.659303+0000 mgr.x (mgr.14152) 574 : cluster [DBG] pgmap v448: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:08.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:07 vm04 bash[20194]: audit 2026-03-21T07:01:07.717746+0000 mgr.x (mgr.14152) 575 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:07 vm04 bash[20194]: audit 2026-03-21T07:01:07.717746+0000 mgr.x (mgr.14152) 575 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:08.475 INFO:teuthology.orchestra.run.vm04.stderr:Note: switching to '0392f78529848ec72469e8e431875cb98d3a5fb4'. 2026-03-21T07:01:08.475 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T07:01:08.475 INFO:teuthology.orchestra.run.vm04.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr:state without impacting any branches by switching back to a branch. 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr: git switch -c 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr:Or undo this operation with: 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr: git switch - 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T07:01:08.476 INFO:teuthology.orchestra.run.vm04.stderr:HEAD is now at 0392f785298 qa/tasks/keystone: restart mariadb for rocky and alma linux too 2026-03-21T07:01:08.485 DEBUG:teuthology.orchestra.run.vm04:> cp -- /home/ubuntu/cephtest/clone.client.1/src/test/cli-integration/rbd/iscsi_client.t /home/ubuntu/cephtest/archive/cram.client.1 2026-03-21T07:01:08.530 DEBUG:teuthology.orchestra.run.vm07:> mkdir -- /home/ubuntu/cephtest/archive/cram.client.2 && python3 -m venv /home/ubuntu/cephtest/virtualenv && /home/ubuntu/cephtest/virtualenv/bin/pip install cram==0.6 2026-03-21T07:01:10.300 INFO:teuthology.orchestra.run.vm07.stdout:Collecting cram==0.6 2026-03-21T07:01:10.301 INFO:teuthology.orchestra.run.vm07.stdout: Using cached cram-0.6-py2.py3-none-any.whl (17 kB) 2026-03-21T07:01:10.312 INFO:teuthology.orchestra.run.vm07.stdout:Installing collected packages: cram 2026-03-21T07:01:10.317 INFO:teuthology.orchestra.run.vm07.stdout:Successfully installed cram-0.6 2026-03-21T07:01:10.345 DEBUG:teuthology.orchestra.run.vm07:> rm -rf /home/ubuntu/cephtest/clone.client.2 && git clone https://github.com/kshtsk/ceph.git /home/ubuntu/cephtest/clone.client.2 && cd /home/ubuntu/cephtest/clone.client.2 && git checkout 0392f78529848ec72469e8e431875cb98d3a5fb4 2026-03-21T07:01:10.349 INFO:teuthology.orchestra.run.vm07.stderr:Cloning into '/home/ubuntu/cephtest/clone.client.2'... 2026-03-21T07:01:11.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:10 vm04 bash[20194]: cluster 2026-03-21T07:01:09.659680+0000 mgr.x (mgr.14152) 576 : cluster [DBG] pgmap v449: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:11.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:10 vm04 bash[20194]: cluster 2026-03-21T07:01:09.659680+0000 mgr.x (mgr.14152) 576 : cluster [DBG] pgmap v449: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:11.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:10 vm02 bash[17657]: cluster 2026-03-21T07:01:09.659680+0000 mgr.x (mgr.14152) 576 : cluster [DBG] pgmap v449: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:11.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:10 vm02 bash[17657]: cluster 2026-03-21T07:01:09.659680+0000 mgr.x (mgr.14152) 576 : cluster [DBG] pgmap v449: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:10 vm07 bash[19945]: cluster 2026-03-21T07:01:09.659680+0000 mgr.x (mgr.14152) 576 : cluster [DBG] pgmap v449: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:11.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:10 vm07 bash[19945]: cluster 2026-03-21T07:01:09.659680+0000 mgr.x (mgr.14152) 576 : cluster [DBG] pgmap v449: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:13.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:12 vm04 bash[20194]: cluster 2026-03-21T07:01:11.659982+0000 mgr.x (mgr.14152) 577 : cluster [DBG] pgmap v450: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:13.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:12 vm04 bash[20194]: cluster 2026-03-21T07:01:11.659982+0000 mgr.x (mgr.14152) 577 : cluster [DBG] pgmap v450: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:13.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:12 vm02 bash[17657]: cluster 2026-03-21T07:01:11.659982+0000 mgr.x (mgr.14152) 577 : cluster [DBG] pgmap v450: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:13.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:12 vm02 bash[17657]: cluster 2026-03-21T07:01:11.659982+0000 mgr.x (mgr.14152) 577 : cluster [DBG] pgmap v450: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:13.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:12 vm07 bash[19945]: cluster 2026-03-21T07:01:11.659982+0000 mgr.x (mgr.14152) 577 : cluster [DBG] pgmap v450: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:13.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:12 vm07 bash[19945]: cluster 2026-03-21T07:01:11.659982+0000 mgr.x (mgr.14152) 577 : cluster [DBG] pgmap v450: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:15.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:14 vm04 bash[20194]: cluster 2026-03-21T07:01:13.660275+0000 mgr.x (mgr.14152) 578 : cluster [DBG] pgmap v451: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:15.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:14 vm04 bash[20194]: cluster 2026-03-21T07:01:13.660275+0000 mgr.x (mgr.14152) 578 : cluster [DBG] pgmap v451: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:15.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:14 vm02 bash[17657]: cluster 2026-03-21T07:01:13.660275+0000 mgr.x (mgr.14152) 578 : cluster [DBG] pgmap v451: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:15.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:14 vm02 bash[17657]: cluster 2026-03-21T07:01:13.660275+0000 mgr.x (mgr.14152) 578 : cluster [DBG] pgmap v451: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:15.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:14 vm07 bash[19945]: cluster 2026-03-21T07:01:13.660275+0000 mgr.x (mgr.14152) 578 : cluster [DBG] pgmap v451: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:15.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:14 vm07 bash[19945]: cluster 2026-03-21T07:01:13.660275+0000 mgr.x (mgr.14152) 578 : cluster [DBG] pgmap v451: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:17.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:16 vm04 bash[20194]: cluster 2026-03-21T07:01:15.660647+0000 mgr.x (mgr.14152) 579 : cluster [DBG] pgmap v452: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:17.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:16 vm04 bash[20194]: cluster 2026-03-21T07:01:15.660647+0000 mgr.x (mgr.14152) 579 : cluster [DBG] pgmap v452: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:17.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:16 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:01:17.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:16 vm02 bash[17657]: cluster 2026-03-21T07:01:15.660647+0000 mgr.x (mgr.14152) 579 : cluster [DBG] pgmap v452: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:17.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:16 vm02 bash[17657]: cluster 2026-03-21T07:01:15.660647+0000 mgr.x (mgr.14152) 579 : cluster [DBG] pgmap v452: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:17.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:16 vm07 bash[19945]: cluster 2026-03-21T07:01:15.660647+0000 mgr.x (mgr.14152) 579 : cluster [DBG] pgmap v452: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:17.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:16 vm07 bash[19945]: cluster 2026-03-21T07:01:15.660647+0000 mgr.x (mgr.14152) 579 : cluster [DBG] pgmap v452: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:18.153 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:17 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:01:19.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:18 vm04 bash[20194]: audit 2026-03-21T07:01:16.892979+0000 mgr.x (mgr.14152) 580 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:19.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:18 vm04 bash[20194]: audit 2026-03-21T07:01:16.892979+0000 mgr.x (mgr.14152) 580 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:19.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:18 vm04 bash[20194]: cluster 2026-03-21T07:01:17.661000+0000 mgr.x (mgr.14152) 581 : cluster [DBG] pgmap v453: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:19.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:18 vm04 bash[20194]: cluster 2026-03-21T07:01:17.661000+0000 mgr.x (mgr.14152) 581 : cluster [DBG] pgmap v453: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:19.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:18 vm04 bash[20194]: audit 2026-03-21T07:01:17.724648+0000 mgr.x (mgr.14152) 582 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:19.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:18 vm04 bash[20194]: audit 2026-03-21T07:01:17.724648+0000 mgr.x (mgr.14152) 582 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:18 vm02 bash[17657]: audit 2026-03-21T07:01:16.892979+0000 mgr.x (mgr.14152) 580 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:18 vm02 bash[17657]: audit 2026-03-21T07:01:16.892979+0000 mgr.x (mgr.14152) 580 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:18 vm02 bash[17657]: cluster 2026-03-21T07:01:17.661000+0000 mgr.x (mgr.14152) 581 : cluster [DBG] pgmap v453: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:18 vm02 bash[17657]: cluster 2026-03-21T07:01:17.661000+0000 mgr.x (mgr.14152) 581 : cluster [DBG] pgmap v453: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:18 vm02 bash[17657]: audit 2026-03-21T07:01:17.724648+0000 mgr.x (mgr.14152) 582 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:19.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:18 vm02 bash[17657]: audit 2026-03-21T07:01:17.724648+0000 mgr.x (mgr.14152) 582 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:19.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:18 vm07 bash[19945]: audit 2026-03-21T07:01:16.892979+0000 mgr.x (mgr.14152) 580 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:19.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:18 vm07 bash[19945]: audit 2026-03-21T07:01:16.892979+0000 mgr.x (mgr.14152) 580 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:19.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:18 vm07 bash[19945]: cluster 2026-03-21T07:01:17.661000+0000 mgr.x (mgr.14152) 581 : cluster [DBG] pgmap v453: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:19.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:18 vm07 bash[19945]: cluster 2026-03-21T07:01:17.661000+0000 mgr.x (mgr.14152) 581 : cluster [DBG] pgmap v453: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:19.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:18 vm07 bash[19945]: audit 2026-03-21T07:01:17.724648+0000 mgr.x (mgr.14152) 582 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:19.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:18 vm07 bash[19945]: audit 2026-03-21T07:01:17.724648+0000 mgr.x (mgr.14152) 582 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:20 vm04 bash[20194]: cluster 2026-03-21T07:01:19.661320+0000 mgr.x (mgr.14152) 583 : cluster [DBG] pgmap v454: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:21.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:20 vm04 bash[20194]: cluster 2026-03-21T07:01:19.661320+0000 mgr.x (mgr.14152) 583 : cluster [DBG] pgmap v454: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:21.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:20 vm02 bash[17657]: cluster 2026-03-21T07:01:19.661320+0000 mgr.x (mgr.14152) 583 : cluster [DBG] pgmap v454: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:21.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:20 vm02 bash[17657]: cluster 2026-03-21T07:01:19.661320+0000 mgr.x (mgr.14152) 583 : cluster [DBG] pgmap v454: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:21.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:20 vm07 bash[19945]: cluster 2026-03-21T07:01:19.661320+0000 mgr.x (mgr.14152) 583 : cluster [DBG] pgmap v454: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:21.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:20 vm07 bash[19945]: cluster 2026-03-21T07:01:19.661320+0000 mgr.x (mgr.14152) 583 : cluster [DBG] pgmap v454: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:23.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:22 vm04 bash[20194]: cluster 2026-03-21T07:01:21.661824+0000 mgr.x (mgr.14152) 584 : cluster [DBG] pgmap v455: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:23.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:22 vm04 bash[20194]: cluster 2026-03-21T07:01:21.661824+0000 mgr.x (mgr.14152) 584 : cluster [DBG] pgmap v455: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:23.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:22 vm02 bash[17657]: cluster 2026-03-21T07:01:21.661824+0000 mgr.x (mgr.14152) 584 : cluster [DBG] pgmap v455: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:23.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:22 vm02 bash[17657]: cluster 2026-03-21T07:01:21.661824+0000 mgr.x (mgr.14152) 584 : cluster [DBG] pgmap v455: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:23.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:22 vm07 bash[19945]: cluster 2026-03-21T07:01:21.661824+0000 mgr.x (mgr.14152) 584 : cluster [DBG] pgmap v455: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:23.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:22 vm07 bash[19945]: cluster 2026-03-21T07:01:21.661824+0000 mgr.x (mgr.14152) 584 : cluster [DBG] pgmap v455: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:25.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:24 vm02 bash[17657]: cluster 2026-03-21T07:01:23.662247+0000 mgr.x (mgr.14152) 585 : cluster [DBG] pgmap v456: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:25.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:24 vm02 bash[17657]: cluster 2026-03-21T07:01:23.662247+0000 mgr.x (mgr.14152) 585 : cluster [DBG] pgmap v456: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:24 vm07 bash[19945]: cluster 2026-03-21T07:01:23.662247+0000 mgr.x (mgr.14152) 585 : cluster [DBG] pgmap v456: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:25.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:24 vm07 bash[19945]: cluster 2026-03-21T07:01:23.662247+0000 mgr.x (mgr.14152) 585 : cluster [DBG] pgmap v456: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:24 vm04 bash[20194]: cluster 2026-03-21T07:01:23.662247+0000 mgr.x (mgr.14152) 585 : cluster [DBG] pgmap v456: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:25.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:24 vm04 bash[20194]: cluster 2026-03-21T07:01:23.662247+0000 mgr.x (mgr.14152) 585 : cluster [DBG] pgmap v456: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:27.138 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:26 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:01:27.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:26 vm02 bash[17657]: cluster 2026-03-21T07:01:25.662610+0000 mgr.x (mgr.14152) 586 : cluster [DBG] pgmap v457: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:27.139 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:26 vm02 bash[17657]: cluster 2026-03-21T07:01:25.662610+0000 mgr.x (mgr.14152) 586 : cluster [DBG] pgmap v457: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:27.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:26 vm07 bash[19945]: cluster 2026-03-21T07:01:25.662610+0000 mgr.x (mgr.14152) 586 : cluster [DBG] pgmap v457: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:27.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:26 vm07 bash[19945]: cluster 2026-03-21T07:01:25.662610+0000 mgr.x (mgr.14152) 586 : cluster [DBG] pgmap v457: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:26 vm04 bash[20194]: cluster 2026-03-21T07:01:25.662610+0000 mgr.x (mgr.14152) 586 : cluster [DBG] pgmap v457: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:27.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:26 vm04 bash[20194]: cluster 2026-03-21T07:01:25.662610+0000 mgr.x (mgr.14152) 586 : cluster [DBG] pgmap v457: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:28.153 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:27 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:26.903662+0000 mgr.x (mgr.14152) 587 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:26.903662+0000 mgr.x (mgr.14152) 587 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: cluster 2026-03-21T07:01:27.663177+0000 mgr.x (mgr.14152) 588 : cluster [DBG] pgmap v458: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: cluster 2026-03-21T07:01:27.663177+0000 mgr.x (mgr.14152) 588 : cluster [DBG] pgmap v458: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:27.735603+0000 mgr.x (mgr.14152) 589 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:27.735603+0000 mgr.x (mgr.14152) 589 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.084552+0000 mon.a (mon.0) 1005 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.084552+0000 mon.a (mon.0) 1005 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.456578+0000 mon.a (mon.0) 1006 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.456578+0000 mon.a (mon.0) 1006 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.456737+0000 mon.a (mon.0) 1007 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.456737+0000 mon.a (mon.0) 1007 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.457692+0000 mon.a (mon.0) 1008 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.457692+0000 mon.a (mon.0) 1008 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.457796+0000 mon.a (mon.0) 1009 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.457796+0000 mon.a (mon.0) 1009 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.458601+0000 mon.a (mon.0) 1010 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.458601+0000 mon.a (mon.0) 1010 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.464238+0000 mon.a (mon.0) 1011 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.464238+0000 mon.a (mon.0) 1011 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.465919+0000 mon.a (mon.0) 1012 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.465919+0000 mon.a (mon.0) 1012 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.466588+0000 mon.a (mon.0) 1013 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.466588+0000 mon.a (mon.0) 1013 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.470397+0000 mon.a (mon.0) 1014 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.470397+0000 mon.a (mon.0) 1014 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.471417+0000 mon.a (mon.0) 1015 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:01:29.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:28 vm02 bash[17657]: audit 2026-03-21T07:01:28.471417+0000 mon.a (mon.0) 1015 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:26.903662+0000 mgr.x (mgr.14152) 587 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:26.903662+0000 mgr.x (mgr.14152) 587 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: cluster 2026-03-21T07:01:27.663177+0000 mgr.x (mgr.14152) 588 : cluster [DBG] pgmap v458: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: cluster 2026-03-21T07:01:27.663177+0000 mgr.x (mgr.14152) 588 : cluster [DBG] pgmap v458: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:27.735603+0000 mgr.x (mgr.14152) 589 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:27.735603+0000 mgr.x (mgr.14152) 589 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.084552+0000 mon.a (mon.0) 1005 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.084552+0000 mon.a (mon.0) 1005 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.456578+0000 mon.a (mon.0) 1006 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.456578+0000 mon.a (mon.0) 1006 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.456737+0000 mon.a (mon.0) 1007 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.456737+0000 mon.a (mon.0) 1007 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.457692+0000 mon.a (mon.0) 1008 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.457692+0000 mon.a (mon.0) 1008 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.457796+0000 mon.a (mon.0) 1009 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.457796+0000 mon.a (mon.0) 1009 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.458601+0000 mon.a (mon.0) 1010 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.458601+0000 mon.a (mon.0) 1010 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.464238+0000 mon.a (mon.0) 1011 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.464238+0000 mon.a (mon.0) 1011 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.465919+0000 mon.a (mon.0) 1012 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.465919+0000 mon.a (mon.0) 1012 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.466588+0000 mon.a (mon.0) 1013 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.466588+0000 mon.a (mon.0) 1013 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.470397+0000 mon.a (mon.0) 1014 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.470397+0000 mon.a (mon.0) 1014 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.471417+0000 mon.a (mon.0) 1015 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:01:29.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:28 vm07 bash[19945]: audit 2026-03-21T07:01:28.471417+0000 mon.a (mon.0) 1015 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:26.903662+0000 mgr.x (mgr.14152) 587 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:26.903662+0000 mgr.x (mgr.14152) 587 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: cluster 2026-03-21T07:01:27.663177+0000 mgr.x (mgr.14152) 588 : cluster [DBG] pgmap v458: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: cluster 2026-03-21T07:01:27.663177+0000 mgr.x (mgr.14152) 588 : cluster [DBG] pgmap v458: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:27.735603+0000 mgr.x (mgr.14152) 589 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:27.735603+0000 mgr.x (mgr.14152) 589 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.084552+0000 mon.a (mon.0) 1005 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.084552+0000 mon.a (mon.0) 1005 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config dump", "format": "json"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.456578+0000 mon.a (mon.0) 1006 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.456578+0000 mon.a (mon.0) 1006 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.0", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.456737+0000 mon.a (mon.0) 1007 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.456737+0000 mon.a (mon.0) 1007 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.3", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.457692+0000 mon.a (mon.0) 1008 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.457692+0000 mon.a (mon.0) 1008 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.2", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.457796+0000 mon.a (mon.0) 1009 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.457796+0000 mon.a (mon.0) 1009 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.1", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.458601+0000 mon.a (mon.0) 1010 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.458601+0000 mon.a (mon.0) 1010 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config rm", "who": "osd.4", "name": "osd_memory_target"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.464238+0000 mon.a (mon.0) 1011 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.464238+0000 mon.a (mon.0) 1011 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.465919+0000 mon.a (mon.0) 1012 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.465919+0000 mon.a (mon.0) 1012 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "config generate-minimal-conf"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.466588+0000 mon.a (mon.0) 1013 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.466588+0000 mon.a (mon.0) 1013 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "auth get", "entity": "client.admin"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.470397+0000 mon.a (mon.0) 1014 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.470397+0000 mon.a (mon.0) 1014 : audit [INF] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.471417+0000 mon.a (mon.0) 1015 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:01:29.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:28 vm04 bash[20194]: audit 2026-03-21T07:01:28.471417+0000 mon.a (mon.0) 1015 : audit [DBG] from='mgr.14152 192.168.123.102:0/4070344376' entity='mgr.x' cmd={"prefix": "osd tree", "states": ["destroyed"], "format": "json"} : dispatch 2026-03-21T07:01:30.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:29 vm02 bash[17657]: cephadm 2026-03-21T07:01:28.458396+0000 mgr.x (mgr.14152) 590 : cephadm [INF] Adjusting osd_memory_target on vm02 to 407.4M 2026-03-21T07:01:30.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:29 vm02 bash[17657]: cephadm 2026-03-21T07:01:28.458396+0000 mgr.x (mgr.14152) 590 : cephadm [INF] Adjusting osd_memory_target on vm02 to 407.4M 2026-03-21T07:01:30.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:29 vm02 bash[17657]: cephadm 2026-03-21T07:01:28.459641+0000 mgr.x (mgr.14152) 591 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 427254988: error parsing value: Value '427254988' is below minimum 939524096 2026-03-21T07:01:30.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:29 vm02 bash[17657]: cephadm 2026-03-21T07:01:28.459641+0000 mgr.x (mgr.14152) 591 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 427254988: error parsing value: Value '427254988' is below minimum 939524096 2026-03-21T07:01:30.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:29 vm02 bash[17657]: cephadm 2026-03-21T07:01:28.459728+0000 mgr.x (mgr.14152) 592 : cephadm [INF] Adjusting osd_memory_target on vm04 to 1978M 2026-03-21T07:01:30.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:29 vm02 bash[17657]: cephadm 2026-03-21T07:01:28.459728+0000 mgr.x (mgr.14152) 592 : cephadm [INF] Adjusting osd_memory_target on vm04 to 1978M 2026-03-21T07:01:30.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:29 vm07 bash[19945]: cephadm 2026-03-21T07:01:28.458396+0000 mgr.x (mgr.14152) 590 : cephadm [INF] Adjusting osd_memory_target on vm02 to 407.4M 2026-03-21T07:01:30.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:29 vm07 bash[19945]: cephadm 2026-03-21T07:01:28.458396+0000 mgr.x (mgr.14152) 590 : cephadm [INF] Adjusting osd_memory_target on vm02 to 407.4M 2026-03-21T07:01:30.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:29 vm07 bash[19945]: cephadm 2026-03-21T07:01:28.459641+0000 mgr.x (mgr.14152) 591 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 427254988: error parsing value: Value '427254988' is below minimum 939524096 2026-03-21T07:01:30.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:29 vm07 bash[19945]: cephadm 2026-03-21T07:01:28.459641+0000 mgr.x (mgr.14152) 591 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 427254988: error parsing value: Value '427254988' is below minimum 939524096 2026-03-21T07:01:30.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:29 vm07 bash[19945]: cephadm 2026-03-21T07:01:28.459728+0000 mgr.x (mgr.14152) 592 : cephadm [INF] Adjusting osd_memory_target on vm04 to 1978M 2026-03-21T07:01:30.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:29 vm07 bash[19945]: cephadm 2026-03-21T07:01:28.459728+0000 mgr.x (mgr.14152) 592 : cephadm [INF] Adjusting osd_memory_target on vm04 to 1978M 2026-03-21T07:01:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:29 vm04 bash[20194]: cephadm 2026-03-21T07:01:28.458396+0000 mgr.x (mgr.14152) 590 : cephadm [INF] Adjusting osd_memory_target on vm02 to 407.4M 2026-03-21T07:01:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:29 vm04 bash[20194]: cephadm 2026-03-21T07:01:28.458396+0000 mgr.x (mgr.14152) 590 : cephadm [INF] Adjusting osd_memory_target on vm02 to 407.4M 2026-03-21T07:01:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:29 vm04 bash[20194]: cephadm 2026-03-21T07:01:28.459641+0000 mgr.x (mgr.14152) 591 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 427254988: error parsing value: Value '427254988' is below minimum 939524096 2026-03-21T07:01:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:29 vm04 bash[20194]: cephadm 2026-03-21T07:01:28.459641+0000 mgr.x (mgr.14152) 591 : cephadm [WRN] Unable to set osd_memory_target on vm02 to 427254988: error parsing value: Value '427254988' is below minimum 939524096 2026-03-21T07:01:30.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:29 vm04 bash[20194]: cephadm 2026-03-21T07:01:28.459728+0000 mgr.x (mgr.14152) 592 : cephadm [INF] Adjusting osd_memory_target on vm04 to 1978M 2026-03-21T07:01:30.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:29 vm04 bash[20194]: cephadm 2026-03-21T07:01:28.459728+0000 mgr.x (mgr.14152) 592 : cephadm [INF] Adjusting osd_memory_target on vm04 to 1978M 2026-03-21T07:01:31.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:30 vm02 bash[17657]: cluster 2026-03-21T07:01:29.663576+0000 mgr.x (mgr.14152) 593 : cluster [DBG] pgmap v459: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:31.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:30 vm02 bash[17657]: cluster 2026-03-21T07:01:29.663576+0000 mgr.x (mgr.14152) 593 : cluster [DBG] pgmap v459: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:31.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:30 vm07 bash[19945]: cluster 2026-03-21T07:01:29.663576+0000 mgr.x (mgr.14152) 593 : cluster [DBG] pgmap v459: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:31.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:30 vm07 bash[19945]: cluster 2026-03-21T07:01:29.663576+0000 mgr.x (mgr.14152) 593 : cluster [DBG] pgmap v459: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:31.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:30 vm04 bash[20194]: cluster 2026-03-21T07:01:29.663576+0000 mgr.x (mgr.14152) 593 : cluster [DBG] pgmap v459: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:31.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:30 vm04 bash[20194]: cluster 2026-03-21T07:01:29.663576+0000 mgr.x (mgr.14152) 593 : cluster [DBG] pgmap v459: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:32.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:31 vm02 bash[17657]: cluster 2026-03-21T07:01:31.663931+0000 mgr.x (mgr.14152) 594 : cluster [DBG] pgmap v460: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:32.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:31 vm02 bash[17657]: cluster 2026-03-21T07:01:31.663931+0000 mgr.x (mgr.14152) 594 : cluster [DBG] pgmap v460: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:31 vm07 bash[19945]: cluster 2026-03-21T07:01:31.663931+0000 mgr.x (mgr.14152) 594 : cluster [DBG] pgmap v460: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:32.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:31 vm07 bash[19945]: cluster 2026-03-21T07:01:31.663931+0000 mgr.x (mgr.14152) 594 : cluster [DBG] pgmap v460: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:31 vm04 bash[20194]: cluster 2026-03-21T07:01:31.663931+0000 mgr.x (mgr.14152) 594 : cluster [DBG] pgmap v460: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:32.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:31 vm04 bash[20194]: cluster 2026-03-21T07:01:31.663931+0000 mgr.x (mgr.14152) 594 : cluster [DBG] pgmap v460: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:35.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:34 vm04 bash[20194]: cluster 2026-03-21T07:01:33.664273+0000 mgr.x (mgr.14152) 595 : cluster [DBG] pgmap v461: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:35.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:34 vm04 bash[20194]: cluster 2026-03-21T07:01:33.664273+0000 mgr.x (mgr.14152) 595 : cluster [DBG] pgmap v461: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:35.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:34 vm02 bash[17657]: cluster 2026-03-21T07:01:33.664273+0000 mgr.x (mgr.14152) 595 : cluster [DBG] pgmap v461: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:35.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:34 vm02 bash[17657]: cluster 2026-03-21T07:01:33.664273+0000 mgr.x (mgr.14152) 595 : cluster [DBG] pgmap v461: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:35.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:34 vm07 bash[19945]: cluster 2026-03-21T07:01:33.664273+0000 mgr.x (mgr.14152) 595 : cluster [DBG] pgmap v461: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:35.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:34 vm07 bash[19945]: cluster 2026-03-21T07:01:33.664273+0000 mgr.x (mgr.14152) 595 : cluster [DBG] pgmap v461: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:36.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:36 vm02 bash[17657]: cluster 2026-03-21T07:01:35.664590+0000 mgr.x (mgr.14152) 596 : cluster [DBG] pgmap v462: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:36.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:36 vm02 bash[17657]: cluster 2026-03-21T07:01:35.664590+0000 mgr.x (mgr.14152) 596 : cluster [DBG] pgmap v462: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:36.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:36 vm04 bash[20194]: cluster 2026-03-21T07:01:35.664590+0000 mgr.x (mgr.14152) 596 : cluster [DBG] pgmap v462: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:36.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:36 vm04 bash[20194]: cluster 2026-03-21T07:01:35.664590+0000 mgr.x (mgr.14152) 596 : cluster [DBG] pgmap v462: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:36.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:36 vm07 bash[19945]: cluster 2026-03-21T07:01:35.664590+0000 mgr.x (mgr.14152) 596 : cluster [DBG] pgmap v462: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:36.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:36 vm07 bash[19945]: cluster 2026-03-21T07:01:35.664590+0000 mgr.x (mgr.14152) 596 : cluster [DBG] pgmap v462: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:37.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:36 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:01:38.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:37 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:01:39.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:38 vm04 bash[20194]: audit 2026-03-21T07:01:36.914403+0000 mgr.x (mgr.14152) 597 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:39.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:38 vm04 bash[20194]: audit 2026-03-21T07:01:36.914403+0000 mgr.x (mgr.14152) 597 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:39.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:38 vm04 bash[20194]: cluster 2026-03-21T07:01:37.664918+0000 mgr.x (mgr.14152) 598 : cluster [DBG] pgmap v463: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:39.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:38 vm04 bash[20194]: cluster 2026-03-21T07:01:37.664918+0000 mgr.x (mgr.14152) 598 : cluster [DBG] pgmap v463: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:39.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:38 vm04 bash[20194]: audit 2026-03-21T07:01:37.745092+0000 mgr.x (mgr.14152) 599 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:39.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:38 vm04 bash[20194]: audit 2026-03-21T07:01:37.745092+0000 mgr.x (mgr.14152) 599 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:39.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:38 vm02 bash[17657]: audit 2026-03-21T07:01:36.914403+0000 mgr.x (mgr.14152) 597 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:39.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:38 vm02 bash[17657]: audit 2026-03-21T07:01:36.914403+0000 mgr.x (mgr.14152) 597 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:39.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:38 vm02 bash[17657]: cluster 2026-03-21T07:01:37.664918+0000 mgr.x (mgr.14152) 598 : cluster [DBG] pgmap v463: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:39.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:38 vm02 bash[17657]: cluster 2026-03-21T07:01:37.664918+0000 mgr.x (mgr.14152) 598 : cluster [DBG] pgmap v463: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:39.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:38 vm02 bash[17657]: audit 2026-03-21T07:01:37.745092+0000 mgr.x (mgr.14152) 599 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:39.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:38 vm02 bash[17657]: audit 2026-03-21T07:01:37.745092+0000 mgr.x (mgr.14152) 599 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:38 vm07 bash[19945]: audit 2026-03-21T07:01:36.914403+0000 mgr.x (mgr.14152) 597 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:38 vm07 bash[19945]: audit 2026-03-21T07:01:36.914403+0000 mgr.x (mgr.14152) 597 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:38 vm07 bash[19945]: cluster 2026-03-21T07:01:37.664918+0000 mgr.x (mgr.14152) 598 : cluster [DBG] pgmap v463: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:38 vm07 bash[19945]: cluster 2026-03-21T07:01:37.664918+0000 mgr.x (mgr.14152) 598 : cluster [DBG] pgmap v463: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:38 vm07 bash[19945]: audit 2026-03-21T07:01:37.745092+0000 mgr.x (mgr.14152) 599 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:39.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:38 vm07 bash[19945]: audit 2026-03-21T07:01:37.745092+0000 mgr.x (mgr.14152) 599 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:40.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:39 vm04 bash[20194]: cluster 2026-03-21T07:01:39.665239+0000 mgr.x (mgr.14152) 600 : cluster [DBG] pgmap v464: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:40.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:39 vm04 bash[20194]: cluster 2026-03-21T07:01:39.665239+0000 mgr.x (mgr.14152) 600 : cluster [DBG] pgmap v464: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:40.386 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:39 vm02 bash[17657]: cluster 2026-03-21T07:01:39.665239+0000 mgr.x (mgr.14152) 600 : cluster [DBG] pgmap v464: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:40.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:39 vm02 bash[17657]: cluster 2026-03-21T07:01:39.665239+0000 mgr.x (mgr.14152) 600 : cluster [DBG] pgmap v464: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:40.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:39 vm07 bash[19945]: cluster 2026-03-21T07:01:39.665239+0000 mgr.x (mgr.14152) 600 : cluster [DBG] pgmap v464: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:40.402 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:39 vm07 bash[19945]: cluster 2026-03-21T07:01:39.665239+0000 mgr.x (mgr.14152) 600 : cluster [DBG] pgmap v464: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:43.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:42 vm04 bash[20194]: cluster 2026-03-21T07:01:41.665596+0000 mgr.x (mgr.14152) 601 : cluster [DBG] pgmap v465: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:43.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:42 vm04 bash[20194]: cluster 2026-03-21T07:01:41.665596+0000 mgr.x (mgr.14152) 601 : cluster [DBG] pgmap v465: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:43.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:42 vm02 bash[17657]: cluster 2026-03-21T07:01:41.665596+0000 mgr.x (mgr.14152) 601 : cluster [DBG] pgmap v465: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:43.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:42 vm02 bash[17657]: cluster 2026-03-21T07:01:41.665596+0000 mgr.x (mgr.14152) 601 : cluster [DBG] pgmap v465: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:43.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:42 vm07 bash[19945]: cluster 2026-03-21T07:01:41.665596+0000 mgr.x (mgr.14152) 601 : cluster [DBG] pgmap v465: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:43.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:42 vm07 bash[19945]: cluster 2026-03-21T07:01:41.665596+0000 mgr.x (mgr.14152) 601 : cluster [DBG] pgmap v465: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:45.009 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:44 vm04 bash[20194]: cluster 2026-03-21T07:01:43.665944+0000 mgr.x (mgr.14152) 602 : cluster [DBG] pgmap v466: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:45.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:44 vm04 bash[20194]: cluster 2026-03-21T07:01:43.665944+0000 mgr.x (mgr.14152) 602 : cluster [DBG] pgmap v466: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:45.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:44 vm02 bash[17657]: cluster 2026-03-21T07:01:43.665944+0000 mgr.x (mgr.14152) 602 : cluster [DBG] pgmap v466: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:45.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:44 vm02 bash[17657]: cluster 2026-03-21T07:01:43.665944+0000 mgr.x (mgr.14152) 602 : cluster [DBG] pgmap v466: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:45.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:44 vm07 bash[19945]: cluster 2026-03-21T07:01:43.665944+0000 mgr.x (mgr.14152) 602 : cluster [DBG] pgmap v466: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:45.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:44 vm07 bash[19945]: cluster 2026-03-21T07:01:43.665944+0000 mgr.x (mgr.14152) 602 : cluster [DBG] pgmap v466: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:47.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:46 vm04 bash[20194]: cluster 2026-03-21T07:01:45.666305+0000 mgr.x (mgr.14152) 603 : cluster [DBG] pgmap v467: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:47.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:46 vm04 bash[20194]: cluster 2026-03-21T07:01:45.666305+0000 mgr.x (mgr.14152) 603 : cluster [DBG] pgmap v467: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:47.136 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:46 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:01:47.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:46 vm02 bash[17657]: cluster 2026-03-21T07:01:45.666305+0000 mgr.x (mgr.14152) 603 : cluster [DBG] pgmap v467: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:47.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:46 vm02 bash[17657]: cluster 2026-03-21T07:01:45.666305+0000 mgr.x (mgr.14152) 603 : cluster [DBG] pgmap v467: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:47.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:46 vm07 bash[19945]: cluster 2026-03-21T07:01:45.666305+0000 mgr.x (mgr.14152) 603 : cluster [DBG] pgmap v467: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:47.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:46 vm07 bash[19945]: cluster 2026-03-21T07:01:45.666305+0000 mgr.x (mgr.14152) 603 : cluster [DBG] pgmap v467: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:48.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:47 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:01:49.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:48 vm04 bash[20194]: audit 2026-03-21T07:01:46.924982+0000 mgr.x (mgr.14152) 604 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:49.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:48 vm04 bash[20194]: audit 2026-03-21T07:01:46.924982+0000 mgr.x (mgr.14152) 604 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:49.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:48 vm04 bash[20194]: cluster 2026-03-21T07:01:47.667238+0000 mgr.x (mgr.14152) 605 : cluster [DBG] pgmap v468: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:49.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:48 vm04 bash[20194]: cluster 2026-03-21T07:01:47.667238+0000 mgr.x (mgr.14152) 605 : cluster [DBG] pgmap v468: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:49.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:48 vm04 bash[20194]: audit 2026-03-21T07:01:47.755767+0000 mgr.x (mgr.14152) 606 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:49.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:48 vm04 bash[20194]: audit 2026-03-21T07:01:47.755767+0000 mgr.x (mgr.14152) 606 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:48 vm02 bash[17657]: audit 2026-03-21T07:01:46.924982+0000 mgr.x (mgr.14152) 604 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:48 vm02 bash[17657]: audit 2026-03-21T07:01:46.924982+0000 mgr.x (mgr.14152) 604 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:48 vm02 bash[17657]: cluster 2026-03-21T07:01:47.667238+0000 mgr.x (mgr.14152) 605 : cluster [DBG] pgmap v468: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:48 vm02 bash[17657]: cluster 2026-03-21T07:01:47.667238+0000 mgr.x (mgr.14152) 605 : cluster [DBG] pgmap v468: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:48 vm02 bash[17657]: audit 2026-03-21T07:01:47.755767+0000 mgr.x (mgr.14152) 606 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:49.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:48 vm02 bash[17657]: audit 2026-03-21T07:01:47.755767+0000 mgr.x (mgr.14152) 606 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:48 vm07 bash[19945]: audit 2026-03-21T07:01:46.924982+0000 mgr.x (mgr.14152) 604 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:48 vm07 bash[19945]: audit 2026-03-21T07:01:46.924982+0000 mgr.x (mgr.14152) 604 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:48 vm07 bash[19945]: cluster 2026-03-21T07:01:47.667238+0000 mgr.x (mgr.14152) 605 : cluster [DBG] pgmap v468: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:48 vm07 bash[19945]: cluster 2026-03-21T07:01:47.667238+0000 mgr.x (mgr.14152) 605 : cluster [DBG] pgmap v468: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:48 vm07 bash[19945]: audit 2026-03-21T07:01:47.755767+0000 mgr.x (mgr.14152) 606 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:49.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:48 vm07 bash[19945]: audit 2026-03-21T07:01:47.755767+0000 mgr.x (mgr.14152) 606 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr:Note: switching to '0392f78529848ec72469e8e431875cb98d3a5fb4'. 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr:You are in 'detached HEAD' state. You can look around, make experimental 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr:changes and commit them, and you can discard any commits you make in this 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr:state without impacting any branches by switching back to a branch. 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr:If you want to create a new branch to retain commits you create, you may 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr:do so (now or later) by using -c with the switch command. Example: 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr: git switch -c 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr:Or undo this operation with: 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr: git switch - 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr:Turn off this advice by setting config variable advice.detachedHead to false 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T07:01:50.438 INFO:teuthology.orchestra.run.vm07.stderr:HEAD is now at 0392f785298 qa/tasks/keystone: restart mariadb for rocky and alma linux too 2026-03-21T07:01:50.446 DEBUG:teuthology.orchestra.run.vm07:> cp -- /home/ubuntu/cephtest/clone.client.2/src/test/cli-integration/rbd/rest_api_delete.t /home/ubuntu/cephtest/archive/cram.client.2 2026-03-21T07:01:50.491 INFO:tasks.cram:Running tests for client.0... 2026-03-21T07:01:50.491 DEBUG:teuthology.orchestra.run.vm02:> CEPH_REF=master CEPH_ID="0" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.0/*.t 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug (LUN.allocate) created datapool/block1 successfully 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block1' to LIO backstore user:rbd 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug (LUN.add_dev_to_lio) Successfully added datapool/block1 to LIO 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug (LUN.allocate) added 'datapool/block1' to LIO and config object 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug _disk update on localhost, successful 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug gateway validation needed for vm02.local 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:50.767 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:50.768 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-21T07:01:50.768 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T07:01:50.768 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug An additional 1 tpg's are required 2026-03-21T07:01:50.768 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T07:01:50.768 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-21T07:01:50.768 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm02.local HTTP/1.1" 200 - 2026-03-21T07:01:50.768 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm02.local HTTP/1.1" 200 - 2026-03-21T07:01:50.768 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug _gateway update on localhost, successful 2026-03-21T07:01:50.768 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm02.local HTTP/1.1" 200 - 2026-03-21T07:01:50.768 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm02.local HTTP/1.1" 200 - 2026-03-21T07:01:50.768 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug gateway validation needed for vm07.local 2026-03-21T07:01:50.792 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:50.792 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:50.848 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug _gateway update on vm07.local, successful 2026-03-21T07:01:50.848 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-21T07:01:50.848 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T07:01:50.848 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug An additional 1 tpg's are required 2026-03-21T07:01:50.848 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T07:01:50.848 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: cluster 2026-03-21T07:01:49.667689+0000 mgr.x (mgr.14152) 607 : cluster [DBG] pgmap v469: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: cluster 2026-03-21T07:01:49.667689+0000 mgr.x (mgr.14152) 607 : cluster [DBG] pgmap v469: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: audit 2026-03-21T07:01:50.609652+0000 mon.c (mon.1) 86 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: audit 2026-03-21T07:01:50.609652+0000 mon.c (mon.1) 86 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: audit 2026-03-21T07:01:50.610400+0000 mon.c (mon.1) 87 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: audit 2026-03-21T07:01:50.610400+0000 mon.c (mon.1) 87 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: audit 2026-03-21T07:01:50.610803+0000 mon.c (mon.1) 88 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: audit 2026-03-21T07:01:50.610803+0000 mon.c (mon.1) 88 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: audit 2026-03-21T07:01:50.618800+0000 mon.c (mon.1) 89 : audit [INF] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: audit 2026-03-21T07:01:50.618800+0000 mon.c (mon.1) 89 : audit [INF] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: audit 2026-03-21T07:01:50.624512+0000 mon.a (mon.0) 1016 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:50.848 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[17657]: audit 2026-03-21T07:01:50.624512+0000 mon.a (mon.0) 1016 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:51.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T07:01:51.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug _gateway update on vm02.local, successful 2026-03-21T07:01:51.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T07:01:51.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:50] "PUT /api/gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T07:01:51.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:50 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block1' to LIO backstore user:rbd 2026-03-21T07:01:51.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug (LUN.add_dev_to_lio) Successfully added datapool/block1 to LIO 2026-03-21T07:01:51.043 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T07:01:51.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug Setup group ano2 for datapool.block1 on tpg 2 (state 1, owner False, failover type 1) 2026-03-21T07:01:51.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug Setup group ao for datapool.block1 on tpg 1 (state 0, owner True, failover type 1) 2026-03-21T07:01:51.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:01:51.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:01:51.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug _targetlun update on vm02.local, successful 2026-03-21T07:01:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: cluster 2026-03-21T07:01:49.667689+0000 mgr.x (mgr.14152) 607 : cluster [DBG] pgmap v469: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:51.151 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: cluster 2026-03-21T07:01:49.667689+0000 mgr.x (mgr.14152) 607 : cluster [DBG] pgmap v469: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: audit 2026-03-21T07:01:50.609652+0000 mon.c (mon.1) 86 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: audit 2026-03-21T07:01:50.609652+0000 mon.c (mon.1) 86 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: audit 2026-03-21T07:01:50.610400+0000 mon.c (mon.1) 87 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: audit 2026-03-21T07:01:50.610400+0000 mon.c (mon.1) 87 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: audit 2026-03-21T07:01:50.610803+0000 mon.c (mon.1) 88 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: audit 2026-03-21T07:01:50.610803+0000 mon.c (mon.1) 88 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: audit 2026-03-21T07:01:50.618800+0000 mon.c (mon.1) 89 : audit [INF] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: audit 2026-03-21T07:01:50.618800+0000 mon.c (mon.1) 89 : audit [INF] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: audit 2026-03-21T07:01:50.624512+0000 mon.a (mon.0) 1016 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:51.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:50 vm07 bash[19945]: audit 2026-03-21T07:01:50.624512+0000 mon.a (mon.0) 1016 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/ip_addresses HTTP/1.1" 200 - 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/checkconf HTTP/1.1" 200 - 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/sysinfo/checkversions HTTP/1.1" 200 - 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: debug Setting up iqn.2003-01.com.redhat.iscsi-gw:ceph-gw 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: debug (Gateway.create_tpg) created TPG '1' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: debug (Gateway.create_tpg) created TPG '2' for target iqn 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: debug (Gateway.create_target) created an iscsi target with iqn of 'iqn.2003-01.com.redhat.iscsi-gw:ceph-gw' 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: debug iqn.2003-01.com.redhat.iscsi-gw:ceph-gw - Adding the IP to the enabled tpg, allowing iSCSI logins 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "PUT /api/_gateway/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/vm07.local HTTP/1.1" 200 - 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:50 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:50] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: cluster 2026-03-21T07:01:49.667689+0000 mgr.x (mgr.14152) 607 : cluster [DBG] pgmap v469: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: cluster 2026-03-21T07:01:49.667689+0000 mgr.x (mgr.14152) 607 : cluster [DBG] pgmap v469: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: audit 2026-03-21T07:01:50.609652+0000 mon.c (mon.1) 86 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: audit 2026-03-21T07:01:50.609652+0000 mon.c (mon.1) 86 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: audit 2026-03-21T07:01:50.610400+0000 mon.c (mon.1) 87 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: audit 2026-03-21T07:01:50.610400+0000 mon.c (mon.1) 87 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: audit 2026-03-21T07:01:50.610803+0000 mon.c (mon.1) 88 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: audit 2026-03-21T07:01:50.610803+0000 mon.c (mon.1) 88 : audit [DBG] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: audit 2026-03-21T07:01:50.618800+0000 mon.c (mon.1) 89 : audit [INF] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: audit 2026-03-21T07:01:50.618800+0000 mon.c (mon.1) 89 : audit [INF] from='client.? 192.168.123.102:0/2900380257' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: audit 2026-03-21T07:01:50.624512+0000 mon.a (mon.0) 1016 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:51.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:50 vm04 bash[20194]: audit 2026-03-21T07:01:50.624512+0000 mon.a (mon.0) 1016 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2538308856"} : dispatch 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug _targetlun update on vm07.local, successful 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "PUT /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "GET /api/config HTTP/1.1" 200 - 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "GET /api/config HTTP/1.1" 200 - 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug _client update on localhost, successful 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug _client update on vm07.local, successful 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:51.302 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "PUT /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:51.385 INFO:tasks.cram.client.0.vm02.stdout:/home/ubuntu/cephtest/archive/cram.client.0/rest_api_create.t: passed 2026-03-21T07:01:51.385 INFO:tasks.cram.client.0.vm02.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-21T07:01:51.391 INFO:tasks.cram:Running tests for client.1... 2026-03-21T07:01:51.391 DEBUG:teuthology.orchestra.run.vm04:> CEPH_REF=master CEPH_ID="1" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.1/*.t 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug (LUN.add_dev_to_lio) Adding image 'datapool/block1' to LIO backstore user:rbd 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug (LUN.add_dev_to_lio) Successfully added datapool/block1 to LIO 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug (LUN.allocate) added block1 to LIO using wwn 'ba21977e-a3c9-467e-89e2-eacfb19cae34' defined by vm02.local 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug Setup group ano2 for datapool.block1 on tpg 2 (state 1, owner False, failover type 1) 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug Setup group ao for datapool.block1 on tpg 1 (state 0, owner True, failover type 1) 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "PUT /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug (Client.define_client) iqn.1994-05.com.redhat:client added successfully 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "PUT /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug (Client.add_lun) added image 'datapool/block1' to iqn.1994-05.com.redhat:client 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: debug ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:51.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:01:51 vm07 bash[42142]: ::ffff:192.168.123.102 - - [21/Mar/2026 07:01:51] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "GET /api/config HTTP/1.1" 200 - 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "GET /api/config HTTP/1.1" 200 - 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug (Client.add_lun) added image 'datapool/block1' to iqn.1994-05.com.redhat:client 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug (main) client 'iqn.1994-05.com.redhat:client' configured without security 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "PUT /api/_clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug _clientlun update on localhost, successful 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug _clientlun update on vm07.local, successful 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:51.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[49158]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:01:51] "PUT /api/clientlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.026447+0000 mon.a (mon.0) 1017 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.026447+0000 mon.a (mon.0) 1017 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.027175+0000 mon.a (mon.0) 1018 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.027175+0000 mon.a (mon.0) 1018 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.027424+0000 mon.a (mon.0) 1019 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.027424+0000 mon.a (mon.0) 1019 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.037026+0000 mon.a (mon.0) 1020 : audit [INF] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2900380257"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.037026+0000 mon.a (mon.0) 1020 : audit [INF] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2900380257"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.159414+0000 mon.c (mon.1) 90 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.159414+0000 mon.c (mon.1) 90 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.160416+0000 mon.c (mon.1) 91 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.160416+0000 mon.c (mon.1) 91 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.160777+0000 mon.c (mon.1) 92 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.160777+0000 mon.c (mon.1) 92 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.171234+0000 mon.c (mon.1) 93 : audit [INF] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.171234+0000 mon.c (mon.1) 93 : audit [INF] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.177222+0000 mon.a (mon.0) 1021 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: audit 2026-03-21T07:01:51.177222+0000 mon.a (mon.0) 1021 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: cluster 2026-03-21T07:01:51.668085+0000 mgr.x (mgr.14152) 608 : cluster [DBG] pgmap v470: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:52.138 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:51 vm02 bash[17657]: cluster 2026-03-21T07:01:51.668085+0000 mgr.x (mgr.14152) 608 : cluster [DBG] pgmap v470: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.026447+0000 mon.a (mon.0) 1017 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.026447+0000 mon.a (mon.0) 1017 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.027175+0000 mon.a (mon.0) 1018 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.027175+0000 mon.a (mon.0) 1018 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.027424+0000 mon.a (mon.0) 1019 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.027424+0000 mon.a (mon.0) 1019 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.037026+0000 mon.a (mon.0) 1020 : audit [INF] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2900380257"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.037026+0000 mon.a (mon.0) 1020 : audit [INF] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2900380257"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.159414+0000 mon.c (mon.1) 90 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.159414+0000 mon.c (mon.1) 90 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.160416+0000 mon.c (mon.1) 91 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.160416+0000 mon.c (mon.1) 91 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.160777+0000 mon.c (mon.1) 92 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.160777+0000 mon.c (mon.1) 92 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.171234+0000 mon.c (mon.1) 93 : audit [INF] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.171234+0000 mon.c (mon.1) 93 : audit [INF] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.177222+0000 mon.a (mon.0) 1021 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: audit 2026-03-21T07:01:51.177222+0000 mon.a (mon.0) 1021 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: cluster 2026-03-21T07:01:51.668085+0000 mgr.x (mgr.14152) 608 : cluster [DBG] pgmap v470: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:52.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:51 vm07 bash[19945]: cluster 2026-03-21T07:01:51.668085+0000 mgr.x (mgr.14152) 608 : cluster [DBG] pgmap v470: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.026447+0000 mon.a (mon.0) 1017 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.026447+0000 mon.a (mon.0) 1017 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.027175+0000 mon.a (mon.0) 1018 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.027175+0000 mon.a (mon.0) 1018 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.027424+0000 mon.a (mon.0) 1019 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.027424+0000 mon.a (mon.0) 1019 : audit [DBG] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.037026+0000 mon.a (mon.0) 1020 : audit [INF] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2900380257"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.037026+0000 mon.a (mon.0) 1020 : audit [INF] from='client.? 192.168.123.102:0/3130816451' entity='client.iscsi.iscsi.a' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.102:0/2900380257"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.159414+0000 mon.c (mon.1) 90 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.159414+0000 mon.c (mon.1) 90 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd pool get", "pool": "datapool", "var": "crush_rule", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.160416+0000 mon.c (mon.1) 91 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.160416+0000 mon.c (mon.1) 91 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "ssd", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.160777+0000 mon.c (mon.1) 92 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.160777+0000 mon.c (mon.1) 92 : audit [DBG] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd crush rule ls-by-class", "class": "nvme", "format": "json"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.171234+0000 mon.c (mon.1) 93 : audit [INF] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.171234+0000 mon.c (mon.1) 93 : audit [INF] from='client.? 192.168.123.107:0/796412880' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.177222+0000 mon.a (mon.0) 1021 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: audit 2026-03-21T07:01:51.177222+0000 mon.a (mon.0) 1021 : audit [INF] from='client.? ' entity='client.iscsi.iscsi.b' cmd={"prefix": "osd blocklist","blocklistop": "rm","addr": "192.168.123.107:0/3792060432"} : dispatch 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: cluster 2026-03-21T07:01:51.668085+0000 mgr.x (mgr.14152) 608 : cluster [DBG] pgmap v470: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:52.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:51 vm04 bash[20194]: cluster 2026-03-21T07:01:51.668085+0000 mgr.x (mgr.14152) 608 : cluster [DBG] pgmap v470: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 2.5 KiB/s rd, 2 op/s 2026-03-21T07:01:55.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:54 vm04 bash[20194]: cluster 2026-03-21T07:01:53.668513+0000 mgr.x (mgr.14152) 609 : cluster [DBG] pgmap v471: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:55.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:54 vm04 bash[20194]: cluster 2026-03-21T07:01:53.668513+0000 mgr.x (mgr.14152) 609 : cluster [DBG] pgmap v471: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:55.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:54 vm02 bash[17657]: cluster 2026-03-21T07:01:53.668513+0000 mgr.x (mgr.14152) 609 : cluster [DBG] pgmap v471: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:55.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:54 vm02 bash[17657]: cluster 2026-03-21T07:01:53.668513+0000 mgr.x (mgr.14152) 609 : cluster [DBG] pgmap v471: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:55.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:54 vm07 bash[19945]: cluster 2026-03-21T07:01:53.668513+0000 mgr.x (mgr.14152) 609 : cluster [DBG] pgmap v471: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:55.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:54 vm07 bash[19945]: cluster 2026-03-21T07:01:53.668513+0000 mgr.x (mgr.14152) 609 : cluster [DBG] pgmap v471: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 1.7 KiB/s rd, 1 op/s 2026-03-21T07:01:57.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:56 vm04 bash[20194]: cluster 2026-03-21T07:01:55.668929+0000 mgr.x (mgr.14152) 610 : cluster [DBG] pgmap v472: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:01:57.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:56 vm04 bash[20194]: cluster 2026-03-21T07:01:55.668929+0000 mgr.x (mgr.14152) 610 : cluster [DBG] pgmap v472: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:01:57.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:56 vm02 bash[17657]: cluster 2026-03-21T07:01:55.668929+0000 mgr.x (mgr.14152) 610 : cluster [DBG] pgmap v472: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:01:57.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:56 vm02 bash[17657]: cluster 2026-03-21T07:01:55.668929+0000 mgr.x (mgr.14152) 610 : cluster [DBG] pgmap v472: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:01:57.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:56 vm07 bash[19945]: cluster 2026-03-21T07:01:55.668929+0000 mgr.x (mgr.14152) 610 : cluster [DBG] pgmap v472: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:01:57.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:56 vm07 bash[19945]: cluster 2026-03-21T07:01:55.668929+0000 mgr.x (mgr.14152) 610 : cluster [DBG] pgmap v472: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:01:59.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:58 vm04 bash[20194]: audit 2026-03-21T07:01:56.935386+0000 mgr.x (mgr.14152) 611 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:59.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:58 vm04 bash[20194]: audit 2026-03-21T07:01:56.935386+0000 mgr.x (mgr.14152) 611 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:59.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:58 vm04 bash[20194]: cluster 2026-03-21T07:01:57.669245+0000 mgr.x (mgr.14152) 612 : cluster [DBG] pgmap v473: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:01:59.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:58 vm04 bash[20194]: cluster 2026-03-21T07:01:57.669245+0000 mgr.x (mgr.14152) 612 : cluster [DBG] pgmap v473: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:01:59.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:58 vm04 bash[20194]: audit 2026-03-21T07:01:57.761361+0000 mgr.x (mgr.14152) 613 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:59.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:01:58 vm04 bash[20194]: audit 2026-03-21T07:01:57.761361+0000 mgr.x (mgr.14152) 613 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:58 vm02 bash[17657]: audit 2026-03-21T07:01:56.935386+0000 mgr.x (mgr.14152) 611 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:58 vm02 bash[17657]: audit 2026-03-21T07:01:56.935386+0000 mgr.x (mgr.14152) 611 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:58 vm02 bash[17657]: cluster 2026-03-21T07:01:57.669245+0000 mgr.x (mgr.14152) 612 : cluster [DBG] pgmap v473: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:01:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:58 vm02 bash[17657]: cluster 2026-03-21T07:01:57.669245+0000 mgr.x (mgr.14152) 612 : cluster [DBG] pgmap v473: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:01:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:58 vm02 bash[17657]: audit 2026-03-21T07:01:57.761361+0000 mgr.x (mgr.14152) 613 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:59.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:01:58 vm02 bash[17657]: audit 2026-03-21T07:01:57.761361+0000 mgr.x (mgr.14152) 613 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:59.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:58 vm07 bash[19945]: audit 2026-03-21T07:01:56.935386+0000 mgr.x (mgr.14152) 611 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:59.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:58 vm07 bash[19945]: audit 2026-03-21T07:01:56.935386+0000 mgr.x (mgr.14152) 611 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:59.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:58 vm07 bash[19945]: cluster 2026-03-21T07:01:57.669245+0000 mgr.x (mgr.14152) 612 : cluster [DBG] pgmap v473: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:01:59.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:58 vm07 bash[19945]: cluster 2026-03-21T07:01:57.669245+0000 mgr.x (mgr.14152) 612 : cluster [DBG] pgmap v473: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:01:59.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:58 vm07 bash[19945]: audit 2026-03-21T07:01:57.761361+0000 mgr.x (mgr.14152) 613 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:01:59.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:01:58 vm07 bash[19945]: audit 2026-03-21T07:01:57.761361+0000 mgr.x (mgr.14152) 613 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:01.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:00 vm04 bash[20194]: cluster 2026-03-21T07:01:59.669636+0000 mgr.x (mgr.14152) 614 : cluster [DBG] pgmap v474: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:01.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:00 vm04 bash[20194]: cluster 2026-03-21T07:01:59.669636+0000 mgr.x (mgr.14152) 614 : cluster [DBG] pgmap v474: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:01.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:00 vm02 bash[17657]: cluster 2026-03-21T07:01:59.669636+0000 mgr.x (mgr.14152) 614 : cluster [DBG] pgmap v474: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:01.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:00 vm02 bash[17657]: cluster 2026-03-21T07:01:59.669636+0000 mgr.x (mgr.14152) 614 : cluster [DBG] pgmap v474: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:01.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:00 vm07 bash[19945]: cluster 2026-03-21T07:01:59.669636+0000 mgr.x (mgr.14152) 614 : cluster [DBG] pgmap v474: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:01.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:00 vm07 bash[19945]: cluster 2026-03-21T07:01:59.669636+0000 mgr.x (mgr.14152) 614 : cluster [DBG] pgmap v474: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:01.694 INFO:tasks.cram.client.1.vm04.stdout:/home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t: passed 2026-03-21T07:02:01.694 INFO:tasks.cram.client.1.vm04.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-21T07:02:01.699 INFO:tasks.cram:Running tests for client.2... 2026-03-21T07:02:01.699 DEBUG:teuthology.orchestra.run.vm07:> CEPH_REF=master CEPH_ID="2" PATH=$PATH:/usr/sbin adjust-ulimits ceph-coverage /home/ubuntu/cephtest/archive/coverage /home/ubuntu/cephtest/virtualenv/bin/cram -v -- /home/ubuntu/cephtest/archive/cram.client.2/*.t 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: debug LUN unmap request received, config commit to be performed by vm07.local 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:01 vm02 bash[49158]: debug Removing target configuration 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:01] "GET /api/config HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:01] "GET /api/config HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug _client update on vm02.local, successful 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug (Client.delete) deleted NodeACL for iqn.1994-05.com.redhat:client 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:01] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:01] "DELETE /api/_client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug _client update on localhost, successful 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:01] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:01] "DELETE /api/client/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw/iqn.1994-05.com.redhat:client HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug _targetlun update on vm02.local, successful 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug LUN unmap request received, config commit to be performed by vm07.local 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "DELETE /api/_targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug _targetlun update on vm07.local, successful 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:01] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:01] "DELETE /api/targetlun/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.007 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.008 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:01 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:01] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.008 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: debug _target update on vm02.local, successful 2026-03-21T07:02:02.184 INFO:tasks.cram.client.2.vm07.stdout:/home/ubuntu/cephtest/archive/cram.client.2/rest_api_delete.t: passed 2026-03-21T07:02:02.184 INFO:tasks.cram.client.2.vm07.stdout:# Ran 1 tests, 0 skipped, 0 failed. 2026-03-21T07:02:02.188 DEBUG:teuthology.orchestra.run.vm02:> test -f /home/ubuntu/cephtest/archive/cram.client.0/rest_api_create.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.0/rest_api_create.t 2026-03-21T07:02:02.191 DEBUG:teuthology.orchestra.run.vm02:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.0 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.0 2026-03-21T07:02:02.386 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:02 vm02 bash[49158]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:02] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:02 vm02 bash[49158]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:02] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:02 vm02 bash[49158]: debug Change detected - internal 24 / xattr 26 refreshing 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: debug (Gateway.load_config) successfully loaded existing target definition 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: debug Removing target configuration 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:02] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:02] "DELETE /api/_target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: debug _target update on vm07.local, successful 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:02] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:02] "DELETE /api/target/iqn.2003-01.com.redhat.iscsi-gw:ceph-gw HTTP/1.1" 200 - 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:02] "GET /api/_ping HTTP/1.1" 200 - 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: debug LUN deletion request received, rbd removal to be performed by vm07.local 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: debug ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:02] "DELETE /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: ::ffff:192.168.123.107 - - [21/Mar/2026 07:02:02] "DELETE /api/_disk/datapool/block1 HTTP/1.1" 200 - 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: debug _disk update on vm07.local, successful 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: debug ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:02] "DELETE /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-21T07:02:02.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:02 vm07 bash[42142]: ::ffff:127.0.0.1 - - [21/Mar/2026 07:02:02] "DELETE /api/disk/datapool/block1 HTTP/1.1" 200 - 2026-03-21T07:02:02.621 DEBUG:teuthology.orchestra.run.vm04:> test -f /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.1/iscsi_client.t 2026-03-21T07:02:02.625 DEBUG:teuthology.orchestra.run.vm04:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.1 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.1 2026-03-21T07:02:03.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:02 vm04 bash[20194]: cluster 2026-03-21T07:02:01.669985+0000 mgr.x (mgr.14152) 615 : cluster [DBG] pgmap v475: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:02:03.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:02 vm04 bash[20194]: cluster 2026-03-21T07:02:01.669985+0000 mgr.x (mgr.14152) 615 : cluster [DBG] pgmap v475: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:02:03.075 DEBUG:teuthology.orchestra.run.vm07:> test -f /home/ubuntu/cephtest/archive/cram.client.2/rest_api_delete.t.err || rm -f -- /home/ubuntu/cephtest/archive/cram.client.2/rest_api_delete.t 2026-03-21T07:02:03.078 DEBUG:teuthology.orchestra.run.vm07:> rm -rf -- /home/ubuntu/cephtest/virtualenv /home/ubuntu/cephtest/clone.client.2 ; rmdir --ignore-fail-on-non-empty /home/ubuntu/cephtest/archive/cram.client.2 2026-03-21T07:02:03.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:02 vm02 bash[17657]: cluster 2026-03-21T07:02:01.669985+0000 mgr.x (mgr.14152) 615 : cluster [DBG] pgmap v475: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:02:03.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:02 vm02 bash[17657]: cluster 2026-03-21T07:02:01.669985+0000 mgr.x (mgr.14152) 615 : cluster [DBG] pgmap v475: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:02:03.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:02 vm07 bash[19945]: cluster 2026-03-21T07:02:01.669985+0000 mgr.x (mgr.14152) 615 : cluster [DBG] pgmap v475: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:02:03.152 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:02 vm07 bash[19945]: cluster 2026-03-21T07:02:01.669985+0000 mgr.x (mgr.14152) 615 : cluster [DBG] pgmap v475: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 34 KiB/s rd, 2.3 KiB/s wr, 37 op/s 2026-03-21T07:02:03.481 DEBUG:teuthology.run_tasks:Unwinding manager ceph_iscsi_client 2026-03-21T07:02:03.483 DEBUG:teuthology.run_tasks:Unwinding manager install 2026-03-21T07:02:03.485 INFO:teuthology.task.install.util:Removing shipped files: /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer... 2026-03-21T07:02:03.485 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-21T07:02:03.487 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-21T07:02:03.487 DEBUG:teuthology.orchestra.run.vm07:> sudo rm -f -- /home/ubuntu/cephtest/valgrind.supp /usr/bin/daemon-helper /usr/bin/adjust-ulimits /usr/bin/stdin-killer 2026-03-21T07:02:03.537 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-21T07:02:03.537 DEBUG:teuthology.orchestra.run.vm02:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-21T07:02:03.542 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-21T07:02:03.542 DEBUG:teuthology.orchestra.run.vm04:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-21T07:02:03.546 INFO:teuthology.task.install.deb:Removing packages: ceph, cephadm, ceph-mds, ceph-mgr, ceph-common, ceph-fuse, ceph-test, ceph-volume, radosgw, python3-rados, python3-rgw, python3-cephfs, python3-rbd, libcephfs2, libcephfs-dev, librados2, librbd1, rbd-fuse on Debian system. 2026-03-21T07:02:03.546 DEBUG:teuthology.orchestra.run.vm07:> for d in ceph cephadm ceph-mds ceph-mgr ceph-common ceph-fuse ceph-test ceph-volume radosgw python3-rados python3-rgw python3-cephfs python3-rbd libcephfs2 libcephfs-dev librados2 librbd1 rbd-fuse ; do sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" purge $d || true ; done 2026-03-21T07:02:03.580 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:03.581 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:03.615 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:03.815 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:03.815 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:03.829 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:03.830 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:03.830 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:03.831 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:03.841 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:03 vm07 bash[19945]: cluster 2026-03-21T07:02:03.670394+0000 mgr.x (mgr.14152) 616 : cluster [DBG] pgmap v476: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:03.841 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:03 vm07 bash[19945]: cluster 2026-03-21T07:02:03.670394+0000 mgr.x (mgr.14152) 616 : cluster [DBG] pgmap v476: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:03 vm04 bash[20194]: cluster 2026-03-21T07:02:03.670394+0000 mgr.x (mgr.14152) 616 : cluster [DBG] pgmap v476: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:04.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:03 vm04 bash[20194]: cluster 2026-03-21T07:02:03.670394+0000 mgr.x (mgr.14152) 616 : cluster [DBG] pgmap v476: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:04.011 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:04.014 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mon libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T07:02:04.014 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:04.031 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:04.031 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be REMOVED: 2026-03-21T07:02:04.033 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:04.034 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T07:02:04.034 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:04.034 INFO:teuthology.orchestra.run.vm07.stdout: ceph* 2026-03-21T07:02:04.036 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T07:02:04.036 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:04.058 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-21T07:02:04.059 INFO:teuthology.orchestra.run.vm02.stdout: ceph* 2026-03-21T07:02:04.063 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be REMOVED: 2026-03-21T07:02:04.065 INFO:teuthology.orchestra.run.vm04.stdout: ceph* 2026-03-21T07:02:04.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:03 vm02 bash[17657]: cluster 2026-03-21T07:02:03.670394+0000 mgr.x (mgr.14152) 616 : cluster [DBG] pgmap v476: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:04.136 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:03 vm02 bash[17657]: cluster 2026-03-21T07:02:03.670394+0000 mgr.x (mgr.14152) 616 : cluster [DBG] pgmap v476: 4 pgs: 4 active+clean; 452 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 33 KiB/s rd, 2.3 KiB/s wr, 36 op/s 2026-03-21T07:02:04.254 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:04.254 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-21T07:02:04.272 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:04.272 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-21T07:02:04.273 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:04.273 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 47.1 kB disk space will be freed. 2026-03-21T07:02:04.292 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126243 files and directories currently installed.) 2026-03-21T07:02:04.294 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:04.312 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126243 files and directories currently installed.) 2026-03-21T07:02:04.312 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126243 files and directories currently installed.) 2026-03-21T07:02:04.314 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:04.315 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:05.669 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:05.713 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:05.756 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:05.756 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:05.797 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:05.798 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:05.970 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:05.971 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:05.991 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:05.992 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:06.005 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:06.005 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:06.138 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:06.139 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T07:02:06.140 INFO:teuthology.orchestra.run.vm02.stdout: python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:06.140 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:06.150 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:06.151 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mon libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T07:02:06.152 INFO:teuthology.orchestra.run.vm07.stdout: python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:06.152 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:06.163 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-21T07:02:06.165 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-cephadm* cephadm* 2026-03-21T07:02:06.171 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be REMOVED: 2026-03-21T07:02:06.172 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-cephadm* cephadm* 2026-03-21T07:02:06.246 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:06.247 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T07:02:06.247 INFO:teuthology.orchestra.run.vm04.stdout: python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:06.247 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:06.261 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be REMOVED: 2026-03-21T07:02:06.263 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-cephadm* cephadm* 2026-03-21T07:02:06.380 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 2 to remove and 36 not upgraded. 2026-03-21T07:02:06.380 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 2177 kB disk space will be freed. 2026-03-21T07:02:06.389 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 2 to remove and 36 not upgraded. 2026-03-21T07:02:06.389 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 2177 kB disk space will be freed. 2026-03-21T07:02:06.433 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126241 files and directories currently installed.) 2026-03-21T07:02:06.435 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:06.437 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126241 files and directories currently installed.) 2026-03-21T07:02:06.440 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:06.454 INFO:teuthology.orchestra.run.vm02.stdout:Removing cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:06.460 INFO:teuthology.orchestra.run.vm07.stdout:Removing cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:06.478 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 2 to remove and 36 not upgraded. 2026-03-21T07:02:06.478 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 2177 kB disk space will be freed. 2026-03-21T07:02:06.486 INFO:teuthology.orchestra.run.vm02.stdout:Looking for files to backup/remove ... 2026-03-21T07:02:06.488 INFO:teuthology.orchestra.run.vm02.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-21T07:02:06.491 INFO:teuthology.orchestra.run.vm02.stdout:Removing user `cephadm' ... 2026-03-21T07:02:06.491 INFO:teuthology.orchestra.run.vm02.stdout:Warning: group `nogroup' has no more members. 2026-03-21T07:02:06.493 INFO:teuthology.orchestra.run.vm07.stdout:Looking for files to backup/remove ... 2026-03-21T07:02:06.495 INFO:teuthology.orchestra.run.vm07.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-21T07:02:06.498 INFO:teuthology.orchestra.run.vm07.stdout:Removing user `cephadm' ... 2026-03-21T07:02:06.498 INFO:teuthology.orchestra.run.vm07.stdout:Warning: group `nogroup' has no more members. 2026-03-21T07:02:06.506 INFO:teuthology.orchestra.run.vm02.stdout:Done. 2026-03-21T07:02:06.512 INFO:teuthology.orchestra.run.vm07.stdout:Done. 2026-03-21T07:02:06.528 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126241 files and directories currently installed.) 2026-03-21T07:02:06.531 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:06.531 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-mgr-cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:06.536 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:06.553 INFO:teuthology.orchestra.run.vm04.stdout:Removing cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:06.586 INFO:teuthology.orchestra.run.vm04.stdout:Looking for files to backup/remove ... 2026-03-21T07:02:06.588 INFO:teuthology.orchestra.run.vm04.stdout:Not backing up/removing `/var/lib/cephadm', it matches ^/var/.*. 2026-03-21T07:02:06.591 INFO:teuthology.orchestra.run.vm04.stdout:Removing user `cephadm' ... 2026-03-21T07:02:06.591 INFO:teuthology.orchestra.run.vm04.stdout:Warning: group `nogroup' has no more members. 2026-03-21T07:02:06.605 INFO:teuthology.orchestra.run.vm04.stdout:Done. 2026-03-21T07:02:06.631 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:06.641 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126155 files and directories currently installed.) 2026-03-21T07:02:06.643 INFO:teuthology.orchestra.run.vm07.stdout:Purging configuration files for cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:06.644 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126155 files and directories currently installed.) 2026-03-21T07:02:06.647 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:06.748 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126155 files and directories currently installed.) 2026-03-21T07:02:06.750 INFO:teuthology.orchestra.run.vm04.stdout:Purging configuration files for cephadm (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:06.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:06 vm02 bash[17657]: cluster 2026-03-21T07:02:05.670875+0000 mgr.x (mgr.14152) 617 : cluster [DBG] pgmap v477: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 99 KiB/s rd, 42 KiB/s wr, 73 op/s 2026-03-21T07:02:06.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:06 vm02 bash[17657]: cluster 2026-03-21T07:02:05.670875+0000 mgr.x (mgr.14152) 617 : cluster [DBG] pgmap v477: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 99 KiB/s rd, 42 KiB/s wr, 73 op/s 2026-03-21T07:02:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:06 vm07 bash[19945]: cluster 2026-03-21T07:02:05.670875+0000 mgr.x (mgr.14152) 617 : cluster [DBG] pgmap v477: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 99 KiB/s rd, 42 KiB/s wr, 73 op/s 2026-03-21T07:02:06.902 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:06 vm07 bash[19945]: cluster 2026-03-21T07:02:05.670875+0000 mgr.x (mgr.14152) 617 : cluster [DBG] pgmap v477: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 99 KiB/s rd, 42 KiB/s wr, 73 op/s 2026-03-21T07:02:07.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:06 vm04 bash[20194]: cluster 2026-03-21T07:02:05.670875+0000 mgr.x (mgr.14152) 617 : cluster [DBG] pgmap v477: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 99 KiB/s rd, 42 KiB/s wr, 73 op/s 2026-03-21T07:02:07.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:06 vm04 bash[20194]: cluster 2026-03-21T07:02:05.670875+0000 mgr.x (mgr.14152) 617 : cluster [DBG] pgmap v477: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 99 KiB/s rd, 42 KiB/s wr, 73 op/s 2026-03-21T07:02:07.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:06 vm02 bash[49158]: debug there is no tcmu-runner data available 2026-03-21T07:02:07.851 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:07.891 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:07.950 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:07.970 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:07.990 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:08.011 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:08.110 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:08.111 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:08.152 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:07 vm07 bash[42142]: debug there is no tcmu-runner data available 2026-03-21T07:02:08.186 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:08.187 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:08.248 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:08.248 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:08.320 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:08.322 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mon libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T07:02:08.322 INFO:teuthology.orchestra.run.vm04.stdout: python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:08.322 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:08.345 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be REMOVED: 2026-03-21T07:02:08.345 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:08.347 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mds* 2026-03-21T07:02:08.347 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mon libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T07:02:08.347 INFO:teuthology.orchestra.run.vm02.stdout: python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:08.347 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:08.366 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-21T07:02:08.368 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mds* 2026-03-21T07:02:08.487 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:08.487 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mon libboost-iostreams1.74.0 libboost-thread1.74.0 libpmemobj1 2026-03-21T07:02:08.487 INFO:teuthology.orchestra.run.vm07.stdout: python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:08.487 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:08.499 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be REMOVED: 2026-03-21T07:02:08.500 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mds* 2026-03-21T07:02:08.558 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:08.558 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 6851 kB disk space will be freed. 2026-03-21T07:02:08.563 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:08.563 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 6851 kB disk space will be freed. 2026-03-21T07:02:08.597 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126155 files and directories currently installed.) 2026-03-21T07:02:08.599 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:08.615 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126155 files and directories currently installed.) 2026-03-21T07:02:08.618 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:08.701 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:08.701 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 6851 kB disk space will be freed. 2026-03-21T07:02:08.749 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126155 files and directories currently installed.) 2026-03-21T07:02:08.752 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:08.876 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:08 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:08.876 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:08 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:08.876 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:08 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:08.876 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:08 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:08.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:08 vm02 bash[17657]: audit 2026-03-21T07:02:06.943698+0000 mgr.x (mgr.14152) 618 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:08.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:08 vm02 bash[17657]: audit 2026-03-21T07:02:06.943698+0000 mgr.x (mgr.14152) 618 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:08.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:08 vm02 bash[17657]: cluster 2026-03-21T07:02:07.671381+0000 mgr.x (mgr.14152) 619 : cluster [DBG] pgmap v478: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:08.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:08 vm02 bash[17657]: cluster 2026-03-21T07:02:07.671381+0000 mgr.x (mgr.14152) 619 : cluster [DBG] pgmap v478: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:08.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:08 vm02 bash[17657]: audit 2026-03-21T07:02:07.769237+0000 mgr.x (mgr.14152) 620 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:08.876 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:08 vm02 bash[17657]: audit 2026-03-21T07:02:07.769237+0000 mgr.x (mgr.14152) 620 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:08.877 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:08 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.000 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:08 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.000 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:08 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.000 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:08 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.000 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:08 vm04 bash[20194]: audit 2026-03-21T07:02:06.943698+0000 mgr.x (mgr.14152) 618 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:09.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:08 vm04 bash[20194]: audit 2026-03-21T07:02:06.943698+0000 mgr.x (mgr.14152) 618 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:09.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:08 vm04 bash[20194]: cluster 2026-03-21T07:02:07.671381+0000 mgr.x (mgr.14152) 619 : cluster [DBG] pgmap v478: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:09.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:08 vm04 bash[20194]: cluster 2026-03-21T07:02:07.671381+0000 mgr.x (mgr.14152) 619 : cluster [DBG] pgmap v478: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:09.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:08 vm04 bash[20194]: audit 2026-03-21T07:02:07.769237+0000 mgr.x (mgr.14152) 620 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:09.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:08 vm04 bash[20194]: audit 2026-03-21T07:02:07.769237+0000 mgr.x (mgr.14152) 620 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:09.001 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:08 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.010 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:08 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.010 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:08 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.011 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:08 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.011 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:08 vm07 bash[19945]: audit 2026-03-21T07:02:06.943698+0000 mgr.x (mgr.14152) 618 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:09.011 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:08 vm07 bash[19945]: audit 2026-03-21T07:02:06.943698+0000 mgr.x (mgr.14152) 618 : audit [DBG] from='client.14490 -' entity='client.iscsi.iscsi.a' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:09.011 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:08 vm07 bash[19945]: cluster 2026-03-21T07:02:07.671381+0000 mgr.x (mgr.14152) 619 : cluster [DBG] pgmap v478: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:09.011 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:08 vm07 bash[19945]: cluster 2026-03-21T07:02:07.671381+0000 mgr.x (mgr.14152) 619 : cluster [DBG] pgmap v478: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:09.011 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:08 vm07 bash[19945]: audit 2026-03-21T07:02:07.769237+0000 mgr.x (mgr.14152) 620 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:09.011 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:08 vm07 bash[19945]: audit 2026-03-21T07:02:07.769237+0000 mgr.x (mgr.14152) 620 : audit [DBG] from='client.24416 -' entity='client.iscsi.iscsi.b' cmd=[{"prefix": "service status", "format": "json"}]: dispatch 2026-03-21T07:02:09.011 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:08 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.011 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:08 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.098 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:09.116 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:09.136 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.137 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.137 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.211 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126147 files and directories currently installed.) 2026-03-21T07:02:09.214 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:09.229 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126147 files and directories currently installed.) 2026-03-21T07:02:09.232 INFO:teuthology.orchestra.run.vm04.stdout:Purging configuration files for ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:09.246 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:09.259 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.260 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.260 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.354 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126147 files and directories currently installed.) 2026-03-21T07:02:09.355 INFO:teuthology.orchestra.run.vm07.stdout:Purging configuration files for ceph-mds (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:09.385 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.386 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.386 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.386 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.386 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.521 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.521 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.521 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.521 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.597 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.598 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.598 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.598 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.598 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.652 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.653 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.653 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.653 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.653 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.887 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.887 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.887 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:09.887 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:09 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:10.010 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:10.010 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:10.010 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:10.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:09 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:10.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:10.153 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:10.153 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:10.153 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:10.153 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:09 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:11.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:10 vm04 bash[20194]: cluster 2026-03-21T07:02:09.671784+0000 mgr.x (mgr.14152) 621 : cluster [DBG] pgmap v479: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:11.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:10 vm04 bash[20194]: cluster 2026-03-21T07:02:09.671784+0000 mgr.x (mgr.14152) 621 : cluster [DBG] pgmap v479: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:11.071 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:11.074 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:10 vm02 bash[17657]: cluster 2026-03-21T07:02:09.671784+0000 mgr.x (mgr.14152) 621 : cluster [DBG] pgmap v479: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:11.074 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:10 vm02 bash[17657]: cluster 2026-03-21T07:02:09.671784+0000 mgr.x (mgr.14152) 621 : cluster [DBG] pgmap v479: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:11.110 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:11.134 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:11.138 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:10 vm07 bash[19945]: cluster 2026-03-21T07:02:09.671784+0000 mgr.x (mgr.14152) 621 : cluster [DBG] pgmap v479: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:11.138 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:10 vm07 bash[19945]: cluster 2026-03-21T07:02:09.671784+0000 mgr.x (mgr.14152) 621 : cluster [DBG] pgmap v479: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:11.160 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:11.177 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:11.202 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:11.382 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:11.383 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:11.411 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:11.411 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:11.454 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:11.455 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:11.536 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:11.537 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core ceph-mon libboost-iostreams1.74.0 2026-03-21T07:02:11.537 INFO:teuthology.orchestra.run.vm02.stdout: libboost-thread1.74.0 libpmemobj1 python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:11.537 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:11.537 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:11.537 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:11.537 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend python3-psutil python3-repoze.lru python3-requests-oauthlib 2026-03-21T07:02:11.537 INFO:teuthology.orchestra.run.vm02.stdout: python3-routes python3-rsa python3-simplejson python3-sklearn 2026-03-21T07:02:11.537 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn-lib python3-tempora python3-threadpoolctl python3-webob 2026-03-21T07:02:11.537 INFO:teuthology.orchestra.run.vm02.stdout: python3-websocket python3-zc.lockfile 2026-03-21T07:02:11.537 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:11.546 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-21T07:02:11.546 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-21T07:02:11.546 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-k8sevents* 2026-03-21T07:02:11.599 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:11.599 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core ceph-mon libboost-iostreams1.74.0 2026-03-21T07:02:11.599 INFO:teuthology.orchestra.run.vm07.stdout: libboost-thread1.74.0 libpmemobj1 python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:11.599 INFO:teuthology.orchestra.run.vm07.stdout: python3-cachetools python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:11.599 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:11.599 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:11.599 INFO:teuthology.orchestra.run.vm07.stdout: python3-portend python3-psutil python3-repoze.lru python3-requests-oauthlib 2026-03-21T07:02:11.599 INFO:teuthology.orchestra.run.vm07.stdout: python3-routes python3-rsa python3-simplejson python3-sklearn 2026-03-21T07:02:11.599 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn-lib python3-tempora python3-threadpoolctl python3-webob 2026-03-21T07:02:11.599 INFO:teuthology.orchestra.run.vm07.stdout: python3-websocket python3-zc.lockfile 2026-03-21T07:02:11.599 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:11.610 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be REMOVED: 2026-03-21T07:02:11.610 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-21T07:02:11.610 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-k8sevents* 2026-03-21T07:02:11.623 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:11.623 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core ceph-mon libboost-iostreams1.74.0 2026-03-21T07:02:11.623 INFO:teuthology.orchestra.run.vm04.stdout: libboost-thread1.74.0 libpmemobj1 python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:11.623 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:11.623 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:11.623 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:11.623 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend python3-psutil python3-repoze.lru python3-requests-oauthlib 2026-03-21T07:02:11.623 INFO:teuthology.orchestra.run.vm04.stdout: python3-routes python3-rsa python3-simplejson python3-sklearn 2026-03-21T07:02:11.623 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn-lib python3-tempora python3-threadpoolctl python3-webob 2026-03-21T07:02:11.623 INFO:teuthology.orchestra.run.vm04.stdout: python3-websocket python3-zc.lockfile 2026-03-21T07:02:11.623 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:11.633 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be REMOVED: 2026-03-21T07:02:11.633 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr* ceph-mgr-dashboard* ceph-mgr-diskprediction-local* 2026-03-21T07:02:11.634 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-k8sevents* 2026-03-21T07:02:11.712 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 4 to remove and 36 not upgraded. 2026-03-21T07:02:11.712 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 219 MB disk space will be freed. 2026-03-21T07:02:11.758 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126147 files and directories currently installed.) 2026-03-21T07:02:11.761 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:11.773 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:11.801 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:11.814 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 4 to remove and 36 not upgraded. 2026-03-21T07:02:11.814 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 219 MB disk space will be freed. 2026-03-21T07:02:11.821 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 4 to remove and 36 not upgraded. 2026-03-21T07:02:11.821 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 219 MB disk space will be freed. 2026-03-21T07:02:11.858 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126147 files and directories currently installed.) 2026-03-21T07:02:11.859 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 126147 files and directories currently installed.) 2026-03-21T07:02:11.860 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:11.861 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-mgr-k8sevents (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:11.872 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:11.874 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-mgr-diskprediction-local (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:11.878 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:11.901 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:11.905 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-mgr-dashboard (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:11.983 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:11.985 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:12.251 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.252 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.252 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.252 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.252 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.259 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.260 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.260 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.260 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.361 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.362 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.362 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.362 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.362 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.431 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 124364 files and directories currently installed.) 2026-03-21T07:02:12.434 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:12.542 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 124364 files and directories currently installed.) 2026-03-21T07:02:12.545 INFO:teuthology.orchestra.run.vm07.stdout:Purging configuration files for ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:12.570 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 124364 files and directories currently installed.) 2026-03-21T07:02:12.573 INFO:teuthology.orchestra.run.vm04.stdout:Purging configuration files for ceph-mgr (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:12.601 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.602 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.602 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.602 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.602 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.603 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.603 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.603 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.604 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.653 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.653 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.653 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.653 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.653 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.858 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.858 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.858 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.858 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.858 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.858 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:12 vm02 bash[17657]: cluster 2026-03-21T07:02:11.672241+0000 mgr.x (mgr.14152) 622 : cluster [DBG] pgmap v480: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:12.859 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:12 vm02 bash[17657]: cluster 2026-03-21T07:02:11.672241+0000 mgr.x (mgr.14152) 622 : cluster [DBG] pgmap v480: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:12.863 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.863 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.863 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:12 vm04 bash[20194]: cluster 2026-03-21T07:02:11.672241+0000 mgr.x (mgr.14152) 622 : cluster [DBG] pgmap v480: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:12.863 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:12 vm04 bash[20194]: cluster 2026-03-21T07:02:11.672241+0000 mgr.x (mgr.14152) 622 : cluster [DBG] pgmap v480: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:12.863 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:12.864 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.137 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.137 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.137 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:12 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:12 vm07 bash[19945]: cluster 2026-03-21T07:02:11.672241+0000 mgr.x (mgr.14152) 622 : cluster [DBG] pgmap v480: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:13.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:12 vm07 bash[19945]: cluster 2026-03-21T07:02:11.672241+0000 mgr.x (mgr.14152) 622 : cluster [DBG] pgmap v480: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:13.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.153 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.153 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.153 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.153 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.153 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.153 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.153 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.153 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:12 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.260 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.260 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.260 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:13.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:12 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:14.184 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:14.220 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:14.263 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:14.273 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:14.298 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:14.311 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:14.409 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:14.409 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:14.451 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:14.451 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:14.498 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:14.499 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:14.565 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:14.565 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:14.565 INFO:teuthology.orchestra.run.vm02.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:14.566 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:14.566 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:14.566 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:14.566 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:14.566 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:14.566 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:14.566 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:14.566 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:14.566 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:14.566 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:14.579 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-21T07:02:14.580 INFO:teuthology.orchestra.run.vm02.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-21T07:02:14.639 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:14.639 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:14.640 INFO:teuthology.orchestra.run.vm04.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:14.640 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:14.640 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:14.640 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:14.640 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:14.640 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:14.640 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:14.640 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:14.640 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:14.640 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:14.640 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:14.651 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be REMOVED: 2026-03-21T07:02:14.652 INFO:teuthology.orchestra.run.vm04.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-21T07:02:14.705 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:14.705 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:14.705 INFO:teuthology.orchestra.run.vm07.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:14.705 INFO:teuthology.orchestra.run.vm07.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:14.705 INFO:teuthology.orchestra.run.vm07.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:14.705 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:14.705 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:14.706 INFO:teuthology.orchestra.run.vm07.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:14.706 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:14.706 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:14.706 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:14.706 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:14.706 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:14.719 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be REMOVED: 2026-03-21T07:02:14.721 INFO:teuthology.orchestra.run.vm07.stdout: ceph-base* ceph-common* ceph-mon* ceph-osd* ceph-test* ceph-volume* radosgw* 2026-03-21T07:02:14.770 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 7 to remove and 36 not upgraded. 2026-03-21T07:02:14.770 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 732 MB disk space will be freed. 2026-03-21T07:02:14.817 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 124364 files and directories currently installed.) 2026-03-21T07:02:14.820 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:14.821 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 7 to remove and 36 not upgraded. 2026-03-21T07:02:14.821 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 732 MB disk space will be freed. 2026-03-21T07:02:14.865 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 124364 files and directories currently installed.) 2026-03-21T07:02:14.868 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:14.883 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:14.921 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 7 to remove and 36 not upgraded. 2026-03-21T07:02:14.921 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 732 MB disk space will be freed. 2026-03-21T07:02:14.936 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:14.965 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 124364 files and directories currently installed.) 2026-03-21T07:02:14.968 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-volume (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:15.035 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:15.044 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:14 vm02 bash[17657]: cluster 2026-03-21T07:02:13.672751+0000 mgr.x (mgr.14152) 623 : cluster [DBG] pgmap v481: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:15.045 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:14 vm02 bash[17657]: cluster 2026-03-21T07:02:13.672751+0000 mgr.x (mgr.14152) 623 : cluster [DBG] pgmap v481: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:15.060 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:14 vm07 bash[19945]: cluster 2026-03-21T07:02:13.672751+0000 mgr.x (mgr.14152) 623 : cluster [DBG] pgmap v481: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:15.060 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:14 vm07 bash[19945]: cluster 2026-03-21T07:02:13.672751+0000 mgr.x (mgr.14152) 623 : cluster [DBG] pgmap v481: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:15.093 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:14 vm04 bash[20194]: cluster 2026-03-21T07:02:13.672751+0000 mgr.x (mgr.14152) 623 : cluster [DBG] pgmap v481: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:15.093 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:14 vm04 bash[20194]: cluster 2026-03-21T07:02:13.672751+0000 mgr.x (mgr.14152) 623 : cluster [DBG] pgmap v481: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 37 op/s 2026-03-21T07:02:15.320 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.320 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.320 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.320 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.320 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.360 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:15.380 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.380 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.380 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.380 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.380 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.380 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.380 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.380 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.380 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.380 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.427 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:15.448 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.448 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.448 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.448 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.448 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.448 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.448 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.448 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.536 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:15.637 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.637 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.637 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.653 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.653 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.653 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.653 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.653 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.760 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.760 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.760 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.801 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:15.880 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:15.943 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.943 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.944 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.944 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.944 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.954 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.955 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.955 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.955 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:15.955 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.042 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.042 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.043 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.043 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:15 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.049 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:16.205 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.206 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.206 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.206 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.206 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:15 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.258 INFO:teuthology.orchestra.run.vm02.stdout:Removing radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:16.279 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.279 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.279 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.279 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.279 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.279 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.279 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.279 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.279 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:15 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.279 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.347 INFO:teuthology.orchestra.run.vm04.stdout:Removing radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:16.367 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.367 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.368 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.368 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.368 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.368 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.368 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.368 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.513 INFO:teuthology.orchestra.run.vm07.stdout:Removing radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:16.533 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.533 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.533 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.533 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.533 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.533 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.533 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.533 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.533 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.533 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.636 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.636 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.636 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.636 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.636 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.722 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.722 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.722 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.722 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.740 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:16.805 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:16.815 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:16.863 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.863 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.863 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.863 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.863 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.863 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:16 vm07 bash[19945]: cluster 2026-03-21T07:02:15.673112+0000 mgr.x (mgr.14152) 624 : cluster [DBG] pgmap v482: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:16.863 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:16 vm07 bash[19945]: cluster 2026-03-21T07:02:15.673112+0000 mgr.x (mgr.14152) 624 : cluster [DBG] pgmap v482: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:16.882 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:16.886 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.886 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: Stopping Ceph mgr.x for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:16.887 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.887 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: Stopping Ceph osd.0 for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:16.887 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.887 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.887 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: Stopping Ceph osd.1 for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:16.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:16 vm02 bash[17657]: cluster 2026-03-21T07:02:15.673112+0000 mgr.x (mgr.14152) 624 : cluster [DBG] pgmap v482: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:16.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:16 vm02 bash[17657]: cluster 2026-03-21T07:02:15.673112+0000 mgr.x (mgr.14152) 624 : cluster [DBG] pgmap v482: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:16.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:16 vm02 systemd[1]: Stopping Ceph mon.a for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:16.969 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-test (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:16.974 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.974 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: Stopping Ceph osd.4 for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:16.974 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.974 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: Stopping Ceph osd.2 for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:16.974 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.974 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: Stopping Ceph osd.3 for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:16.974 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:16.974 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:16 vm04 bash[20194]: cluster 2026-03-21T07:02:15.673112+0000 mgr.x (mgr.14152) 624 : cluster [DBG] pgmap v482: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:16.974 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:16 vm04 bash[20194]: cluster 2026-03-21T07:02:15.673112+0000 mgr.x (mgr.14152) 624 : cluster [DBG] pgmap v482: 4 pgs: 4 active+clean; 449 KiB data, 219 MiB used, 160 GiB / 160 GiB avail; 67 KiB/s rd, 39 KiB/s wr, 38 op/s 2026-03-21T07:02:16.974 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:16 vm04 systemd[1]: Stopping Ceph mon.b for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:17.046 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:17.152 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.153 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: Stopping Ceph osd.5 for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:17.153 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.153 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: Stopping Ceph osd.7 for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:17.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: Stopping Ceph mon.c for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:17.153 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.153 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: Stopping Ceph osd.6 for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:17.153 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:16 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.231 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:16 vm02 bash[17657]: debug 2026-03-21T07:02:16.936+0000 7fe930c2a640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.a -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-21T07:02:17.231 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:16 vm02 bash[17657]: debug 2026-03-21T07:02:16.936+0000 7fe930c2a640 -1 mon.a@0(leader) e3 *** Got Signal Terminated *** 2026-03-21T07:02:17.231 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.231 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.231 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:16 vm02 bash[30274]: debug 2026-03-21T07:02:16.928+0000 7f8aa4e94640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.0 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-21T07:02:17.231 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:16 vm02 bash[30274]: debug 2026-03-21T07:02:16.928+0000 7f8aa4e94640 -1 osd.0 69 *** Got signal Terminated *** 2026-03-21T07:02:17.231 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:16 vm02 bash[30274]: debug 2026-03-21T07:02:16.928+0000 7f8aa4e94640 -1 osd.0 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-21T07:02:17.231 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.231 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:16 vm02 bash[38374]: debug 2026-03-21T07:02:16.924+0000 7f702bc23640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.1 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-21T07:02:17.231 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:16 vm02 bash[38374]: debug 2026-03-21T07:02:16.924+0000 7f702bc23640 -1 osd.1 69 *** Got signal Terminated *** 2026-03-21T07:02:17.231 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:16 vm02 bash[38374]: debug 2026-03-21T07:02:16.924+0000 7f702bc23640 -1 osd.1 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-21T07:02:17.232 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.232 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:16 vm04 bash[20194]: debug 2026-03-21T07:02:16.965+0000 7f1bd86bd640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.b -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-21T07:02:17.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:16 vm04 bash[20194]: debug 2026-03-21T07:02:16.965+0000 7f1bd86bd640 -1 mon.b@2(peon) e3 *** Got Signal Terminated *** 2026-03-21T07:02:17.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:16 vm04 bash[33334]: debug 2026-03-21T07:02:16.981+0000 7f692d9ec640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.4 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:16 vm04 bash[33334]: debug 2026-03-21T07:02:16.981+0000 7f692d9ec640 -1 osd.4 69 *** Got signal Terminated *** 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:16 vm04 bash[33334]: debug 2026-03-21T07:02:16.981+0000 7f692d9ec640 -1 osd.4 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:16 vm04 bash[23182]: debug 2026-03-21T07:02:16.977+0000 7f3b5a14c640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.2 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:16 vm04 bash[23182]: debug 2026-03-21T07:02:16.977+0000 7f3b5a14c640 -1 osd.2 69 *** Got signal Terminated *** 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:16 vm04 bash[23182]: debug 2026-03-21T07:02:16.977+0000 7f3b5a14c640 -1 osd.2 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:16 vm04 bash[29705]: debug 2026-03-21T07:02:16.965+0000 7f38caafc640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.3 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:16 vm04 bash[29705]: debug 2026-03-21T07:02:16.965+0000 7f38caafc640 -1 osd.3 69 *** Got signal Terminated *** 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:16 vm04 bash[29705]: debug 2026-03-21T07:02:16.965+0000 7f38caafc640 -1 osd.3 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-21T07:02:17.260 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.364 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:17.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:17 vm07 bash[19945]: debug 2026-03-21T07:02:17.189+0000 7f7b6fe6f640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-mon -n mon.c -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --default-mon-cluster-log-to-file=false --default-mon-cluster-log-to-stderr=true (PID: 1) UID: 0 2026-03-21T07:02:17.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:17 vm07 bash[19945]: debug 2026-03-21T07:02:17.189+0000 7f7b6fe6f640 -1 mon.c@1(peon) e3 *** Got Signal Terminated *** 2026-03-21T07:02:17.450 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:17 vm07 bash[35243]: debug 2026-03-21T07:02:17.181+0000 7fefa9559640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.6 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:17 vm07 bash[35243]: debug 2026-03-21T07:02:17.181+0000 7fefa9559640 -1 osd.6 69 *** Got signal Terminated *** 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:17 vm07 bash[35243]: debug 2026-03-21T07:02:17.181+0000 7fefa9559640 -1 osd.6 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.451 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:17 vm07 bash[37964]: debug 2026-03-21T07:02:17.181+0000 7f4d1ce9c640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.7 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:17 vm07 bash[37964]: debug 2026-03-21T07:02:17.181+0000 7f4d1ce9c640 -1 osd.7 69 *** Got signal Terminated *** 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:17 vm07 bash[37964]: debug 2026-03-21T07:02:17.181+0000 7f4d1ce9c640 -1 osd.7 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:17 vm07 bash[25201]: debug 2026-03-21T07:02:17.173+0000 7f8a46ab3640 -1 received signal: Terminated from /sbin/docker-init -- /usr/bin/ceph-osd -n osd.5 -f --setuser ceph --setgroup ceph --default-log-to-file=false --default-log-to-stderr=true --default-log-stderr-prefix=debug --osd-objectstore=bluestore (PID: 1) UID: 0 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:17 vm07 bash[25201]: debug 2026-03-21T07:02:17.173+0000 7f8a46ab3640 -1 osd.5 69 *** Got signal Terminated *** 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:17 vm07 bash[25201]: debug 2026-03-21T07:02:17.173+0000 7f8a46ab3640 -1 osd.5 69 *** Immediate shutdown (osd_fast_shutdown=true) *** 2026-03-21T07:02:17.451 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.459 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:17.466 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:17.495 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:17.507 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.508 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:17 vm02 bash[61606]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-mon-a 2026-03-21T07:02:17.508 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.a.service: Deactivated successfully. 2026-03-21T07:02:17.508 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: Stopped Ceph mon.a for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:17.508 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.508 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.508 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: Stopping Ceph iscsi.iscsi.a for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:17.508 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:17 vm02 bash[49158]: debug Shutdown received 2026-03-21T07:02:17.508 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:17 vm02 bash[49158]: debug No gateway configuration to remove on this host (vm02.local) 2026-03-21T07:02:17.508 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.508 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:17 vm02 bash[61573]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-mgr-x 2026-03-21T07:02:17.508 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mgr.x.service: Deactivated successfully. 2026-03-21T07:02:17.508 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: Stopped Ceph mgr.x for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:17.508 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.576 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123871 files and directories currently installed.) 2026-03-21T07:02:17.580 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:17.585 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:17.587 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123871 files and directories currently installed.) 2026-03-21T07:02:17.590 INFO:teuthology.orchestra.run.vm04.stdout:Purging configuration files for radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:17.602 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.602 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:17 vm04 bash[44636]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-mon-b 2026-03-21T07:02:17.602 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.602 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.b.service: Deactivated successfully. 2026-03-21T07:02:17.602 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: Stopped Ceph mon.b for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:17.602 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.602 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.637 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:17.704 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:17 vm07 bash[57991]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-mon-c 2026-03-21T07:02:17.704 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.704 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.c.service: Deactivated successfully. 2026-03-21T07:02:17.704 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: Stopped Ceph mon.c for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:17.704 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.705 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.705 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.705 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: Stopping Ceph iscsi.iscsi.b for b16ecafc-24f1-11f1-8ede-8330751617ee... 2026-03-21T07:02:17.705 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:17 vm07 bash[42142]: debug Shutdown received 2026-03-21T07:02:17.705 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:17 vm07 bash[42142]: debug No gateway configuration to remove on this host (vm07.local) 2026-03-21T07:02:17.705 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.738 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123871 files and directories currently installed.) 2026-03-21T07:02:17.740 INFO:teuthology.orchestra.run.vm07.stdout:Purging configuration files for radosgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:17.797 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.797 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.797 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:17 vm02 bash[61881]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-iscsi-iscsi-a 2026-03-21T07:02:17.797 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.797 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.797 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.940 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.940 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.941 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.941 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.992 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.993 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.993 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.993 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:17 vm07 bash[58221]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-iscsi-iscsi-b 2026-03-21T07:02:17.993 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:17.993 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:17 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.137 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@iscsi.iscsi.a.service: Deactivated successfully. 2026-03-21T07:02:18.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: Stopped Ceph iscsi.iscsi.a for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:18.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.137 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.137 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:17 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.237 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:18.260 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.260 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.260 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.260 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.260 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.260 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:17 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.260 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.261 INFO:teuthology.orchestra.run.vm04.stdout:Purging configuration files for ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:18.341 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.341 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.341 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.341 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.341 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@iscsi.iscsi.b.service: Deactivated successfully. 2026-03-21T07:02:18.341 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: Stopped Ceph iscsi.iscsi.b for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:18.341 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.399 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.399 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.399 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.400 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.400 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.433 INFO:teuthology.orchestra.run.vm07.stdout:Purging configuration files for ceph-mon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:18.644 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.644 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.644 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.644 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.652 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.653 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.653 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.653 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.653 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.653 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.653 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.653 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.653 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.653 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.707 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:18.740 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.740 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.740 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.740 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.742 INFO:teuthology.orchestra.run.vm04.stdout:Purging configuration files for ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:18.742 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.742 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.742 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.742 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.742 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.742 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.880 INFO:teuthology.orchestra.run.vm07.stdout:Purging configuration files for ceph-base (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:18.901 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.901 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.901 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:18.901 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.042 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.042 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.042 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.043 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.043 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:18 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.064 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.064 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.064 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.064 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.065 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:18 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.151 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:19.221 INFO:teuthology.orchestra.run.vm04.stdout:Purging configuration files for ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:19.232 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.232 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.232 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.232 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.232 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.232 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.232 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:18 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.232 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.320 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.320 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.320 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.320 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.320 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.320 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.355 INFO:teuthology.orchestra.run.vm07.stdout:Purging configuration files for ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:19.368 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.368 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.368 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.368 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.368 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.368 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.368 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.368 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.368 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.368 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.502 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.502 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.502 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.502 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.608 INFO:teuthology.orchestra.run.vm02.stdout:dpkg: warning: while removing ceph-common, directory '/var/lib/ceph' not empty so not removed 2026-03-21T07:02:19.619 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:19.637 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.637 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.637 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.637 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.637 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.653 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.653 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.653 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.653 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.653 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.718 INFO:teuthology.orchestra.run.vm04.stdout:dpkg: warning: while removing ceph-common, directory '/var/lib/ceph' not empty so not removed 2026-03-21T07:02:19.728 INFO:teuthology.orchestra.run.vm04.stdout:Purging configuration files for ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:19.758 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.759 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.759 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.759 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:19.815 INFO:teuthology.orchestra.run.vm07.stdout:dpkg: warning: while removing ceph-common, directory '/var/lib/ceph' not empty so not removed 2026-03-21T07:02:19.825 INFO:teuthology.orchestra.run.vm07.stdout:Purging configuration files for ceph-osd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:20.003 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.003 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.003 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.003 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.003 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:19 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.010 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.010 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.010 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:19 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.059 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.059 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.059 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.059 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.059 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.059 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.059 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.059 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.059 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.059 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:19 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:20 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.387 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:20 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:20 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.387 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:20 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.387 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:20 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.402 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:20 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:20 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.403 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:20 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.403 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:20 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:20 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.510 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:20 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.510 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:20 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.510 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:20 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:20.511 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:20 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:21.420 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:21.448 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:21.456 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:21.485 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:21.522 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:21.562 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:21.653 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:21.654 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:21.689 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:21.689 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:21.765 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:21.766 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:21.855 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:21.855 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:21.856 INFO:teuthology.orchestra.run.vm02.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:21.856 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:21.856 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:21.856 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:21.856 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:21.856 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:21.856 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:21.856 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:21.856 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:21.856 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:21.856 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:21.877 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-21T07:02:21.879 INFO:teuthology.orchestra.run.vm02.stdout: ceph-fuse* 2026-03-21T07:02:21.948 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:21.948 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:21.949 INFO:teuthology.orchestra.run.vm04.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:21.949 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:21.949 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:21.949 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:21.949 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:21.949 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:21.949 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:21.949 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:21.949 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:21.949 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:21.949 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:21.966 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be REMOVED: 2026-03-21T07:02:21.968 INFO:teuthology.orchestra.run.vm04.stdout: ceph-fuse* 2026-03-21T07:02:22.063 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:22.064 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:22.073 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be REMOVED: 2026-03-21T07:02:22.073 INFO:teuthology.orchestra.run.vm07.stdout: ceph-fuse* 2026-03-21T07:02:22.221 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:22.221 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 2932 kB disk space will be freed. 2026-03-21T07:02:22.251 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:22.251 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 2932 kB disk space will be freed. 2026-03-21T07:02:22.276 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:22.277 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 2932 kB disk space will be freed. 2026-03-21T07:02:22.290 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123857 files and directories currently installed.) 2026-03-21T07:02:22.294 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:22.308 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:22 vm02 bash[61604]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-0 2026-03-21T07:02:22.308 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:22 vm02 bash[61603]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-1 2026-03-21T07:02:22.310 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123857 files and directories currently installed.) 2026-03-21T07:02:22.314 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:22.384 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:22 vm04 bash[44650]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-2 2026-03-21T07:02:22.384 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:22 vm04 bash[44632]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-3 2026-03-21T07:02:22.385 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:22 vm04 bash[44638]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-4 2026-03-21T07:02:22.431 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123857 files and directories currently installed.) 2026-03-21T07:02:22.434 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:22.563 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.563 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.564 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.564 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.564 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.595 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:22 vm07 bash[57939]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-5 2026-03-21T07:02:22.595 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:22 vm07 bash[57969]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-6 2026-03-21T07:02:22.595 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:22 vm07 bash[57980]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee-osd-7 2026-03-21T07:02:22.757 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.758 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.758 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.758 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.831 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:22.850 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.851 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.851 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.851 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.851 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.859 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.859 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.859 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.859 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.859 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:22.864 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:22.951 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123848 files and directories currently installed.) 2026-03-21T07:02:22.954 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:22.970 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:22.972 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123848 files and directories currently installed.) 2026-03-21T07:02:22.974 INFO:teuthology.orchestra.run.vm04.stdout:Purging configuration files for ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:23.010 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.010 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.2.service: Deactivated successfully. 2026-03-21T07:02:23.010 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: Stopped Ceph osd.2 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:23.010 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.010 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.4.service: Deactivated successfully. 2026-03-21T07:02:23.010 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: Stopped Ceph osd.4 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:23.010 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.010 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.3.service: Deactivated successfully. 2026-03-21T07:02:23.010 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: Stopped Ceph osd.3 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:23.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:22 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.097 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123848 files and directories currently installed.) 2026-03-21T07:02:23.100 INFO:teuthology.orchestra.run.vm07.stdout:Purging configuration files for ceph-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:23.129 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.129 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.5.service: Deactivated successfully. 2026-03-21T07:02:23.130 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: Stopped Ceph osd.5 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:23.130 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.130 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.6.service: Deactivated successfully. 2026-03-21T07:02:23.130 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: Stopped Ceph osd.6 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:23.130 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.130 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.7.service: Deactivated successfully. 2026-03-21T07:02:23.130 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: Stopped Ceph osd.7 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:23.130 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.130 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:22 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.137 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:23 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.137 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:23 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.137 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.0.service: Deactivated successfully. 2026-03-21T07:02:23.137 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: Stopped Ceph osd.0 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:23.137 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:23 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.137 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.1.service: Deactivated successfully. 2026-03-21T07:02:23.137 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:22 vm02 systemd[1]: Stopped Ceph osd.1 for b16ecafc-24f1-11f1-8ede-8330751617ee. 2026-03-21T07:02:23.137 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:23 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.137 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:23 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.403 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.403 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.403 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.511 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:23 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.511 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:23 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.511 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:23 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.511 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:23 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.511 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:23 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.511 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:23 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.511 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:23 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.511 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:23 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.637 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:23 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.637 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:23 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.637 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:23 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.637 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:23 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.637 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:23 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.903 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.903 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.903 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:23.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:23 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:24.711 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:24.751 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:24.813 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:24.852 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:24.945 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:24.964 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:24.964 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:24.987 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:25.050 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:25.050 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:25.144 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:25.176 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:25.176 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:25.213 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:25.239 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:25.240 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:25.253 INFO:teuthology.orchestra.run.vm02.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-21T07:02:25.253 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:25.253 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:25.254 INFO:teuthology.orchestra.run.vm02.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:25.254 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:25.254 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:25.254 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:25.254 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:25.254 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:25.254 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:25.254 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:25.254 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:25.254 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:25.254 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:25.288 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:25.289 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:25.328 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout:Package 'ceph-test' is not installed, so not removed 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:25.458 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:25.459 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:25.476 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:25.476 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:25.485 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:25.486 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:25.509 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:25.553 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:25.553 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:25.628 INFO:teuthology.orchestra.run.vm04.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-21T07:02:25.628 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:25.628 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:25.629 INFO:teuthology.orchestra.run.vm04.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:25.629 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:25.629 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:25.629 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:25.629 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:25.629 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:25.629 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:25.629 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:25.629 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:25.629 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:25.629 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:25.650 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:25.650 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:25.677 INFO:teuthology.orchestra.run.vm02.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-21T07:02:25.677 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:25.677 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:25.678 INFO:teuthology.orchestra.run.vm02.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:25.678 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:25.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:25.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:25.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:25.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:25.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:25.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:25.678 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:25.678 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:25.678 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:25.683 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:25.699 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:25.699 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:25.705 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:25.705 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:25.734 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout:Package 'ceph-volume' is not installed, so not removed 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:25.839 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:25.855 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:25.855 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:25.873 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:25.874 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:25.888 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:25.918 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:25.918 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:26.013 INFO:teuthology.orchestra.run.vm04.stdout:Package 'radosgw' is not installed, so not removed 2026-03-21T07:02:26.013 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:26.013 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:26.013 INFO:teuthology.orchestra.run.vm04.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:26.014 INFO:teuthology.orchestra.run.vm04.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:26.014 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:26.014 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:26.014 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:26.014 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:26.014 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:26.014 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:26.014 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:26.014 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:26.014 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:26.033 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:26.033 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:26.051 INFO:teuthology.orchestra.run.vm02.stdout:Package 'radosgw' is not installed, so not removed 2026-03-21T07:02:26.051 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:26.051 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:26.052 INFO:teuthology.orchestra.run.vm02.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:26.052 INFO:teuthology.orchestra.run.vm02.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:26.052 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:26.052 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:26.052 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:26.052 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:26.052 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:26.052 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:26.052 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:26.052 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:26.052 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:26.067 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:26.069 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:26.069 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:26.081 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:26.082 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:26.102 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout:Package 'radosgw' is not installed, so not removed 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 libsqlite3-mod-ceph 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout: nvme-cli python-asyncssh-doc python3-asyncssh python3-cachetools 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:26.205 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:26.222 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:26.222 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:26.249 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:26.249 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:26.254 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:26.287 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:26.287 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:26.384 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:26.384 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:26.384 INFO:teuthology.orchestra.run.vm04.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:26.385 INFO:teuthology.orchestra.run.vm04.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:26.385 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:26.385 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:26.385 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:26.385 INFO:teuthology.orchestra.run.vm04.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:26.385 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:26.385 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:26.385 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:26.385 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:26.385 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:26.385 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:26.396 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be REMOVED: 2026-03-21T07:02:26.396 INFO:teuthology.orchestra.run.vm04.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-21T07:02:26.425 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:26.425 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:26.425 INFO:teuthology.orchestra.run.vm02.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:26.426 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:26.426 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:26.426 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:26.426 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:26.426 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:26.426 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:26.426 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:26.426 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:26.426 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:26.426 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:26.426 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:26.435 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-21T07:02:26.435 INFO:teuthology.orchestra.run.vm02.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-21T07:02:26.447 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:26.447 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:26.560 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 3 to remove and 36 not upgraded. 2026-03-21T07:02:26.560 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 2086 kB disk space will be freed. 2026-03-21T07:02:26.565 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:26.565 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:26.565 INFO:teuthology.orchestra.run.vm07.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:26.565 INFO:teuthology.orchestra.run.vm07.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:26.565 INFO:teuthology.orchestra.run.vm07.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:26.565 INFO:teuthology.orchestra.run.vm07.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:26.565 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:26.565 INFO:teuthology.orchestra.run.vm07.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:26.565 INFO:teuthology.orchestra.run.vm07.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:26.565 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:26.565 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:26.566 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:26.566 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:26.566 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:26.600 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be REMOVED: 2026-03-21T07:02:26.600 INFO:teuthology.orchestra.run.vm07.stdout: python3-cephfs* python3-rados* python3-rgw* 2026-03-21T07:02:26.622 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 3 to remove and 36 not upgraded. 2026-03-21T07:02:26.622 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 2086 kB disk space will be freed. 2026-03-21T07:02:26.626 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123848 files and directories currently installed.) 2026-03-21T07:02:26.628 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:26.668 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123848 files and directories currently installed.) 2026-03-21T07:02:26.670 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:26.670 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:26.705 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:26.707 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:26.719 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:26.823 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 3 to remove and 36 not upgraded. 2026-03-21T07:02:26.823 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 2086 kB disk space will be freed. 2026-03-21T07:02:26.870 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123848 files and directories currently installed.) 2026-03-21T07:02:26.873 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-cephfs (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:26.888 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-rgw (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:26.902 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-rados (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:27.963 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:28.003 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:28.190 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:28.227 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:28.261 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:28.262 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:28.410 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:28.451 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:28.501 INFO:teuthology.orchestra.run.vm04.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-21T07:02:28.501 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:28.501 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:28.501 INFO:teuthology.orchestra.run.vm04.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:28.502 INFO:teuthology.orchestra.run.vm04.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:28.502 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:28.502 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:28.502 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:28.502 INFO:teuthology.orchestra.run.vm04.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:28.502 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:28.502 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:28.502 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:28.502 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:28.503 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:28.503 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:28.503 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:28.504 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:28.541 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:28.541 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:28.578 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:28.650 INFO:teuthology.orchestra.run.vm02.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-21T07:02:28.650 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:28.650 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:28.650 INFO:teuthology.orchestra.run.vm02.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:28.651 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:28.651 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:28.651 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:28.651 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:28.651 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:28.651 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:28.651 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:28.651 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:28.651 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:28.651 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:28.651 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:28.679 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:28.679 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:28.680 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:28.681 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:28.725 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:28.728 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:28.729 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:28.818 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:28.819 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:28.819 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:28.819 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:28.840 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:28.840 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:28.874 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout:Package 'python3-rgw' is not installed, so not removed 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:28.876 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:28.893 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:28.893 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:28.915 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:28.916 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:28.926 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:29.030 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:29.047 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:29.047 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:29.065 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:29.065 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:29.079 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:29.109 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:29.109 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:29.192 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:29.192 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:29.192 INFO:teuthology.orchestra.run.vm04.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:29.192 INFO:teuthology.orchestra.run.vm04.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:29.192 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:29.192 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:29.192 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:29.192 INFO:teuthology.orchestra.run.vm04.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:29.192 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:29.192 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:29.192 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:29.193 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:29.193 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:29.193 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:29.203 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be REMOVED: 2026-03-21T07:02:29.204 INFO:teuthology.orchestra.run.vm04.stdout: python3-rbd* 2026-03-21T07:02:29.244 INFO:teuthology.orchestra.run.vm07.stdout:Package 'python3-cephfs' is not installed, so not removed 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:29.245 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:29.261 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:29.261 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:29.270 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:29.270 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:29.294 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:29.373 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:29.373 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 1205 kB disk space will be freed. 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:29.386 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:29.396 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-21T07:02:29.396 INFO:teuthology.orchestra.run.vm02.stdout: python3-rbd* 2026-03-21T07:02:29.414 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123824 files and directories currently installed.) 2026-03-21T07:02:29.416 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:29.468 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:29.469 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:29.556 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:29.556 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 1205 kB disk space will be freed. 2026-03-21T07:02:29.598 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123824 files and directories currently installed.) 2026-03-21T07:02:29.600 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 librdkafka1 librgw2 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: libsqlite3-mod-ceph nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:29.605 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:29.615 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be REMOVED: 2026-03-21T07:02:29.615 INFO:teuthology.orchestra.run.vm07.stdout: python3-rbd* 2026-03-21T07:02:29.774 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 1 to remove and 36 not upgraded. 2026-03-21T07:02:29.774 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 1205 kB disk space will be freed. 2026-03-21T07:02:29.809 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123824 files and directories currently installed.) 2026-03-21T07:02:29.811 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-rbd (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:30.663 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:30.700 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:30.933 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:30.934 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:31.040 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:31.079 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:31.118 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:31.118 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:31.118 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-proxy2 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-21T07:02:31.119 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka1 librgw2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-21T07:02:31.119 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-21T07:02:31.119 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:31.119 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:31.119 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:31.119 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:31.119 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:31.119 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:31.119 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:31.119 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:31.119 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:31.133 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be REMOVED: 2026-03-21T07:02:31.134 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-daemon* libcephfs-dev* libcephfs2* 2026-03-21T07:02:31.181 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:31.219 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:31.309 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:31.310 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:31.329 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 3 to remove and 36 not upgraded. 2026-03-21T07:02:31.329 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 2851 kB disk space will be freed. 2026-03-21T07:02:31.374 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123816 files and directories currently installed.) 2026-03-21T07:02:31.377 INFO:teuthology.orchestra.run.vm04.stdout:Removing libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:31.409 INFO:teuthology.orchestra.run.vm04.stdout:Removing libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:31.415 INFO:teuthology.orchestra.run.vm04.stdout:dpkg: warning: while removing libcephfs-dev, directory '/usr/lib/pkgconfig' not empty so not removed 2026-03-21T07:02:31.421 INFO:teuthology.orchestra.run.vm04.stdout:Removing libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:31.447 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:31.451 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:31.452 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:31.461 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:31.462 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:31.462 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-proxy2 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-21T07:02:31.463 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka1 librgw2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-21T07:02:31.463 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-21T07:02:31.463 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:31.463 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:31.463 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:31.463 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:31.463 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:31.463 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:31.463 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:31.463 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:31.463 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:31.480 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-21T07:02:31.482 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-daemon* libcephfs-dev* libcephfs2* 2026-03-21T07:02:31.689 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 3 to remove and 36 not upgraded. 2026-03-21T07:02:31.689 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 2851 kB disk space will be freed. 2026-03-21T07:02:31.729 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123816 files and directories currently installed.) 2026-03-21T07:02:31.730 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:31.730 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:31.731 INFO:teuthology.orchestra.run.vm07.stdout: libcephfs-proxy2 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm07.stdout: librdkafka1 librgw2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm07.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm07.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm07.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:31.732 INFO:teuthology.orchestra.run.vm02.stdout:Removing libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:31.744 INFO:teuthology.orchestra.run.vm02.stdout:Removing libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:31.748 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be REMOVED: 2026-03-21T07:02:31.749 INFO:teuthology.orchestra.run.vm07.stdout: libcephfs-daemon* libcephfs-dev* libcephfs2* 2026-03-21T07:02:31.749 INFO:teuthology.orchestra.run.vm02.stdout:dpkg: warning: while removing libcephfs-dev, directory '/usr/lib/pkgconfig' not empty so not removed 2026-03-21T07:02:31.754 INFO:teuthology.orchestra.run.vm02.stdout:Removing libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:31.782 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:31.939 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 3 to remove and 36 not upgraded. 2026-03-21T07:02:31.939 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 2851 kB disk space will be freed. 2026-03-21T07:02:31.988 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123816 files and directories currently installed.) 2026-03-21T07:02:31.991 INFO:teuthology.orchestra.run.vm07.stdout:Removing libcephfs-daemon (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:32.003 INFO:teuthology.orchestra.run.vm07.stdout:Removing libcephfs-dev (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:32.009 INFO:teuthology.orchestra.run.vm07.stdout:dpkg: warning: while removing libcephfs-dev, directory '/usr/lib/pkgconfig' not empty so not removed 2026-03-21T07:02:32.015 INFO:teuthology.orchestra.run.vm07.stdout:Removing libcephfs2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:32.042 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:32.896 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:32.932 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:33.156 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:33.157 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:33.217 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:33.257 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-proxy2 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka1 librgw2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:33.360 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:33.377 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:33.378 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:33.413 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:33.471 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:33.487 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:33.488 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:33.510 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:33.622 INFO:teuthology.orchestra.run.vm02.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-21T07:02:33.622 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:33.622 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:33.622 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-proxy2 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-21T07:02:33.623 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka1 librgw2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-21T07:02:33.623 INFO:teuthology.orchestra.run.vm02.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-21T07:02:33.623 INFO:teuthology.orchestra.run.vm02.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:33.623 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:33.623 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:33.623 INFO:teuthology.orchestra.run.vm02.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:33.623 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:33.623 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:33.623 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:33.623 INFO:teuthology.orchestra.run.vm02.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:33.623 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:33.641 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:33.641 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:33.647 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:33.647 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:33.684 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:33.716 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:33.716 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:33.866 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:33.866 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:33.866 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:33.866 INFO:teuthology.orchestra.run.vm04.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:33.866 INFO:teuthology.orchestra.run.vm04.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:33.867 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:33.867 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:33.867 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:33.867 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:33.867 INFO:teuthology.orchestra.run.vm04.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:33.867 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:33.867 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:33.867 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:33.867 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:33.867 INFO:teuthology.orchestra.run.vm04.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:33.867 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:33.883 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be REMOVED: 2026-03-21T07:02:33.884 INFO:teuthology.orchestra.run.vm04.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-21T07:02:33.885 INFO:teuthology.orchestra.run.vm04.stdout: qemu-block-extra* rbd-fuse* 2026-03-21T07:02:33.949 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:33.950 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:33.961 INFO:teuthology.orchestra.run.vm07.stdout:Package 'libcephfs-dev' is not installed, so not removed 2026-03-21T07:02:33.961 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:33.961 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:33.962 INFO:teuthology.orchestra.run.vm07.stdout: libcephfs-proxy2 libjq1 liboath0 libonig5 libpmemobj1 libradosstriper1 2026-03-21T07:02:33.963 INFO:teuthology.orchestra.run.vm07.stdout: librdkafka1 librgw2 libsqlite3-mod-ceph nvme-cli python-asyncssh-doc 2026-03-21T07:02:33.963 INFO:teuthology.orchestra.run.vm07.stdout: python3-asyncssh python3-cachetools python3-ceph-argparse 2026-03-21T07:02:33.963 INFO:teuthology.orchestra.run.vm07.stdout: python3-ceph-common python3-cheroot python3-cherrypy3 python3-google-auth 2026-03-21T07:02:33.963 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.classes python3-jaraco.collections python3-jaraco.functools 2026-03-21T07:02:33.963 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.text python3-joblib python3-kubernetes python3-natsort 2026-03-21T07:02:33.963 INFO:teuthology.orchestra.run.vm07.stdout: python3-portend python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:33.963 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:33.963 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:33.963 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:33.963 INFO:teuthology.orchestra.run.vm07.stdout: smartmontools socat xmlstarlet 2026-03-21T07:02:33.963 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:34.005 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:34.006 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:34.042 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:34.112 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 7 to remove and 36 not upgraded. 2026-03-21T07:02:34.113 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 59.2 MB disk space will be freed. 2026-03-21T07:02:34.157 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123794 files and directories currently installed.) 2026-03-21T07:02:34.159 INFO:teuthology.orchestra.run.vm04.stdout:Removing rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.173 INFO:teuthology.orchestra.run.vm04.stdout:Removing libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:34.182 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:34.187 INFO:teuthology.orchestra.run.vm04.stdout:Removing libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.191 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-21T07:02:34.191 INFO:teuthology.orchestra.run.vm02.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-21T07:02:34.191 INFO:teuthology.orchestra.run.vm02.stdout: qemu-block-extra* rbd-fuse* 2026-03-21T07:02:34.201 INFO:teuthology.orchestra.run.vm04.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-21T07:02:34.255 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:34.255 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:34.382 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 7 to remove and 36 not upgraded. 2026-03-21T07:02:34.382 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 59.2 MB disk space will be freed. 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:34.413 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:34.421 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123794 files and directories currently installed.) 2026-03-21T07:02:34.422 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be REMOVED: 2026-03-21T07:02:34.422 INFO:teuthology.orchestra.run.vm07.stdout: librados2* libradosstriper1* librbd1* librgw2* libsqlite3-mod-ceph* 2026-03-21T07:02:34.422 INFO:teuthology.orchestra.run.vm07.stdout: qemu-block-extra* rbd-fuse* 2026-03-21T07:02:34.423 INFO:teuthology.orchestra.run.vm02.stdout:Removing rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.437 INFO:teuthology.orchestra.run.vm02.stdout:Removing libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.450 INFO:teuthology.orchestra.run.vm02.stdout:Removing libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.462 INFO:teuthology.orchestra.run.vm02.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-21T07:02:34.510 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.510 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.510 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.510 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.607 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 7 to remove and 36 not upgraded. 2026-03-21T07:02:34.607 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 59.2 MB disk space will be freed. 2026-03-21T07:02:34.633 INFO:teuthology.orchestra.run.vm04.stdout:Removing librbd1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.647 INFO:teuthology.orchestra.run.vm04.stdout:Removing librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.648 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123794 files and directories currently installed.) 2026-03-21T07:02:34.651 INFO:teuthology.orchestra.run.vm07.stdout:Removing rbd-fuse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.662 INFO:teuthology.orchestra.run.vm04.stdout:Removing librados2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.663 INFO:teuthology.orchestra.run.vm07.stdout:Removing libsqlite3-mod-ceph (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.676 INFO:teuthology.orchestra.run.vm07.stdout:Removing libradosstriper1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.688 INFO:teuthology.orchestra.run.vm07.stdout:Removing qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-21T07:02:34.690 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:34.724 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:34.799 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123743 files and directories currently installed.) 2026-03-21T07:02:34.801 INFO:teuthology.orchestra.run.vm04.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-21T07:02:34.807 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:34 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.807 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:34 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.807 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:34 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.807 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:34 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.807 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:34 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.828 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.828 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.828 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.828 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:34.894 INFO:teuthology.orchestra.run.vm02.stdout:Removing librbd1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.909 INFO:teuthology.orchestra.run.vm02.stdout:Removing librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.927 INFO:teuthology.orchestra.run.vm02.stdout:Removing librados2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:34.955 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:34.995 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:35.032 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:34 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.033 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:34 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.033 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:34 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.033 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:34 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.033 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:34 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.081 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123743 files and directories currently installed.) 2026-03-21T07:02:35.084 INFO:teuthology.orchestra.run.vm02.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-21T07:02:35.116 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:34 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.116 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:34 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.116 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:34 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.116 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:34 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.116 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:34 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.122 INFO:teuthology.orchestra.run.vm07.stdout:Removing librbd1 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:35.136 INFO:teuthology.orchestra.run.vm07.stdout:Removing librgw2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:35.152 INFO:teuthology.orchestra.run.vm07.stdout:Removing librados2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:35.161 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.161 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.161 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.161 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:34 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.180 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:35.219 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:35.309 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123743 files and directories currently installed.) 2026-03-21T07:02:35.312 INFO:teuthology.orchestra.run.vm07.stdout:Purging configuration files for qemu-block-extra (1:6.2+dfsg-2ubuntu6.28) ... 2026-03-21T07:02:35.347 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.347 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.347 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.347 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.347 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.387 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:35 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.387 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:35 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:35 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:35 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.387 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:35 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.510 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:35 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.510 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:35 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.510 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:35 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.510 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:35 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.652 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.653 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.653 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.653 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.653 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:35 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.887 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:35 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.887 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:35 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.887 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:35 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.887 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:35 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.902 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.903 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:35.903 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:35 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:36.741 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:36.774 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:36.776 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:36.816 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:36.958 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:36.959 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:37.000 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:37.000 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:37.086 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:37.095 INFO:teuthology.orchestra.run.vm04.stdout:Package 'librbd1' is not installed, so not removed 2026-03-21T07:02:37.095 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:37.095 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:37.095 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:37.095 INFO:teuthology.orchestra.run.vm04.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:37.095 INFO:teuthology.orchestra.run.vm04.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:37.096 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:37.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:37.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:37.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:37.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:37.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:37.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:37.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:37.096 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:37.096 INFO:teuthology.orchestra.run.vm04.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:37.096 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:37.116 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:37.116 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:37.120 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout:Package 'librbd1' is not installed, so not removed 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:37.147 INFO:teuthology.orchestra.run.vm02.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:37.148 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:37.148 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:37.169 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:37.169 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:37.201 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:37.298 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:37.298 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:37.324 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:37.325 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:37.382 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:37.382 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:37.433 INFO:teuthology.orchestra.run.vm07.stdout:Package 'librbd1' is not installed, so not removed 2026-03-21T07:02:37.433 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:37.433 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:37.433 INFO:teuthology.orchestra.run.vm07.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:37.433 INFO:teuthology.orchestra.run.vm07.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:37.433 INFO:teuthology.orchestra.run.vm07.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:37.434 INFO:teuthology.orchestra.run.vm07.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:37.434 INFO:teuthology.orchestra.run.vm07.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:37.434 INFO:teuthology.orchestra.run.vm07.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:37.434 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:37.434 INFO:teuthology.orchestra.run.vm07.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:37.434 INFO:teuthology.orchestra.run.vm07.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:37.434 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:37.434 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:37.434 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:37.434 INFO:teuthology.orchestra.run.vm07.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:37.434 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:37.456 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:37.457 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:37.489 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:37.491 INFO:teuthology.orchestra.run.vm04.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-21T07:02:37.491 INFO:teuthology.orchestra.run.vm04.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:37.492 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:37.492 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:37.492 INFO:teuthology.orchestra.run.vm04.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:37.492 INFO:teuthology.orchestra.run.vm04.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:37.493 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:37.493 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:37.493 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:37.493 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:37.493 INFO:teuthology.orchestra.run.vm04.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:37.493 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:37.493 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:37.493 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:37.493 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:37.493 INFO:teuthology.orchestra.run.vm04.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:37.493 INFO:teuthology.orchestra.run.vm04.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:37.515 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:37.515 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:37.517 DEBUG:teuthology.orchestra.run.vm04:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:37.530 INFO:teuthology.orchestra.run.vm02.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:37.545 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:37.545 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:37.547 DEBUG:teuthology.orchestra.run.vm02:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-21T07:02:37.570 DEBUG:teuthology.orchestra.run.vm04:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-21T07:02:37.603 DEBUG:teuthology.orchestra.run.vm02:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-21T07:02:37.644 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:37.662 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:37.662 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:37.681 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout:Package 'rbd-fuse' is not installed, so not removed 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout:The following packages were automatically installed and are no longer required: 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:37.862 INFO:teuthology.orchestra.run.vm07.stdout:Use 'sudo apt autoremove' to remove them. 2026-03-21T07:02:37.867 INFO:teuthology.orchestra.run.vm04.stdout:Building dependency tree... 2026-03-21T07:02:37.867 INFO:teuthology.orchestra.run.vm04.stdout:Reading state information... 2026-03-21T07:02:37.891 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 0 to remove and 36 not upgraded. 2026-03-21T07:02:37.891 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:37.893 DEBUG:teuthology.orchestra.run.vm07:> dpkg -l | grep '^.\(U\|H\)R' | awk '{print $2}' | sudo xargs --no-run-if-empty dpkg -P --force-remove-reinstreq 2026-03-21T07:02:37.919 INFO:teuthology.orchestra.run.vm02.stdout:Building dependency tree... 2026-03-21T07:02:37.919 INFO:teuthology.orchestra.run.vm02.stdout:Reading state information... 2026-03-21T07:02:37.953 DEBUG:teuthology.orchestra.run.vm07:> sudo DEBIAN_FRONTEND=noninteractive apt-get -y --force-yes -o Dpkg::Options::="--force-confdef" -o Dpkg::Options::="--force-confold" autoremove 2026-03-21T07:02:38.041 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:38.127 INFO:teuthology.orchestra.run.vm04.stdout:The following packages will be REMOVED: 2026-03-21T07:02:38.127 INFO:teuthology.orchestra.run.vm04.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:38.127 INFO:teuthology.orchestra.run.vm04.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:38.127 INFO:teuthology.orchestra.run.vm04.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:38.127 INFO:teuthology.orchestra.run.vm04.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:38.128 INFO:teuthology.orchestra.run.vm04.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:38.128 INFO:teuthology.orchestra.run.vm04.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:38.128 INFO:teuthology.orchestra.run.vm04.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:38.128 INFO:teuthology.orchestra.run.vm04.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:38.128 INFO:teuthology.orchestra.run.vm04.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:38.128 INFO:teuthology.orchestra.run.vm04.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:38.128 INFO:teuthology.orchestra.run.vm04.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:38.128 INFO:teuthology.orchestra.run.vm04.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:38.128 INFO:teuthology.orchestra.run.vm04.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:38.128 INFO:teuthology.orchestra.run.vm04.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:38.193 INFO:teuthology.orchestra.run.vm02.stdout:The following packages will be REMOVED: 2026-03-21T07:02:38.193 INFO:teuthology.orchestra.run.vm02.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:38.193 INFO:teuthology.orchestra.run.vm02.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:38.194 INFO:teuthology.orchestra.run.vm02.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:38.194 INFO:teuthology.orchestra.run.vm02.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:38.194 INFO:teuthology.orchestra.run.vm02.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:38.194 INFO:teuthology.orchestra.run.vm02.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:38.194 INFO:teuthology.orchestra.run.vm02.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:38.195 INFO:teuthology.orchestra.run.vm02.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:38.195 INFO:teuthology.orchestra.run.vm02.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:38.195 INFO:teuthology.orchestra.run.vm02.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:38.195 INFO:teuthology.orchestra.run.vm02.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:38.195 INFO:teuthology.orchestra.run.vm02.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:38.195 INFO:teuthology.orchestra.run.vm02.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:38.195 INFO:teuthology.orchestra.run.vm02.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:38.309 INFO:teuthology.orchestra.run.vm07.stdout:Building dependency tree... 2026-03-21T07:02:38.310 INFO:teuthology.orchestra.run.vm07.stdout:Reading state information... 2026-03-21T07:02:38.336 INFO:teuthology.orchestra.run.vm04.stdout:0 upgraded, 0 newly installed, 60 to remove and 36 not upgraded. 2026-03-21T07:02:38.336 INFO:teuthology.orchestra.run.vm04.stdout:After this operation, 93.5 MB disk space will be freed. 2026-03-21T07:02:38.384 INFO:teuthology.orchestra.run.vm04.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123743 files and directories currently installed.) 2026-03-21T07:02:38.387 INFO:teuthology.orchestra.run.vm04.stdout:Removing ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:38.395 INFO:teuthology.orchestra.run.vm02.stdout:0 upgraded, 0 newly installed, 60 to remove and 36 not upgraded. 2026-03-21T07:02:38.395 INFO:teuthology.orchestra.run.vm02.stdout:After this operation, 93.5 MB disk space will be freed. 2026-03-21T07:02:38.403 INFO:teuthology.orchestra.run.vm04.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-21T07:02:38.416 INFO:teuthology.orchestra.run.vm04.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-21T07:02:38.429 INFO:teuthology.orchestra.run.vm04.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-21T07:02:38.441 INFO:teuthology.orchestra.run.vm04.stdout:Removing libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:38.449 INFO:teuthology.orchestra.run.vm02.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123743 files and directories currently installed.) 2026-03-21T07:02:38.452 INFO:teuthology.orchestra.run.vm02.stdout:Removing ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:38.454 INFO:teuthology.orchestra.run.vm04.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-21T07:02:38.465 INFO:teuthology.orchestra.run.vm04.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T07:02:38.467 INFO:teuthology.orchestra.run.vm02.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-21T07:02:38.478 INFO:teuthology.orchestra.run.vm04.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T07:02:38.479 INFO:teuthology.orchestra.run.vm02.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-21T07:02:38.491 INFO:teuthology.orchestra.run.vm04.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T07:02:38.491 INFO:teuthology.orchestra.run.vm02.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-21T07:02:38.502 INFO:teuthology.orchestra.run.vm02.stdout:Removing libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:38.512 INFO:teuthology.orchestra.run.vm02.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-21T07:02:38.516 INFO:teuthology.orchestra.run.vm04.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-21T07:02:38.523 INFO:teuthology.orchestra.run.vm02.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T07:02:38.528 INFO:teuthology.orchestra.run.vm04.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-21T07:02:38.534 INFO:teuthology.orchestra.run.vm02.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T07:02:38.541 INFO:teuthology.orchestra.run.vm04.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.545 INFO:teuthology.orchestra.run.vm02.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T07:02:38.553 INFO:teuthology.orchestra.run.vm04.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout:The following packages will be REMOVED: 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: ceph-mgr-modules-core jq libboost-iostreams1.74.0 libboost-thread1.74.0 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: libcephfs-proxy2 libdouble-conversion3 libfuse2 libgfapi0 libgfrpc0 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: libgfxdr0 libglusterfs0 libiscsi7 libjq1 liblttng-ust1 libnbd0 liboath0 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: libonig5 libpcre2-16-0 libpmemobj1 libqt5core5a libqt5dbus5 libqt5network5 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: librdkafka1 libthrift-0.16.0 nvme-cli python-asyncssh-doc python3-asyncssh 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: python3-cachetools python3-ceph-argparse python3-ceph-common python3-cheroot 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: python3-cherrypy3 python3-google-auth python3-jaraco.classes 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: python3-jaraco.collections python3-jaraco.functools python3-jaraco.text 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: python3-joblib python3-kubernetes python3-natsort python3-portend 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: python3-prettytable python3-psutil python3-repoze.lru 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: python3-requests-oauthlib python3-routes python3-rsa python3-simplejson 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: python3-sklearn python3-sklearn-lib python3-tempora python3-threadpoolctl 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: python3-wcwidth python3-webob python3-websocket python3-zc.lockfile 2026-03-21T07:02:38.557 INFO:teuthology.orchestra.run.vm07.stdout: qttranslations5-l10n smartmontools socat xmlstarlet 2026-03-21T07:02:38.566 INFO:teuthology.orchestra.run.vm02.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-21T07:02:38.566 INFO:teuthology.orchestra.run.vm04.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.577 INFO:teuthology.orchestra.run.vm02.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-21T07:02:38.581 INFO:teuthology.orchestra.run.vm04.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.587 INFO:teuthology.orchestra.run.vm02.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.594 INFO:teuthology.orchestra.run.vm04.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-21T07:02:38.599 INFO:teuthology.orchestra.run.vm02.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.608 INFO:teuthology.orchestra.run.vm04.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-21T07:02:38.611 INFO:teuthology.orchestra.run.vm02.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.623 INFO:teuthology.orchestra.run.vm02.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.623 INFO:teuthology.orchestra.run.vm04.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-21T07:02:38.635 INFO:teuthology.orchestra.run.vm02.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-21T07:02:38.638 INFO:teuthology.orchestra.run.vm04.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-21T07:02:38.647 INFO:teuthology.orchestra.run.vm02.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-21T07:02:38.652 INFO:teuthology.orchestra.run.vm04.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-21T07:02:38.658 INFO:teuthology.orchestra.run.vm02.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-21T07:02:38.665 INFO:teuthology.orchestra.run.vm04.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-21T07:02:38.670 INFO:teuthology.orchestra.run.vm02.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-21T07:02:38.679 INFO:teuthology.orchestra.run.vm04.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-21T07:02:38.680 INFO:teuthology.orchestra.run.vm02.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-21T07:02:38.690 INFO:teuthology.orchestra.run.vm02.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-21T07:02:38.691 INFO:teuthology.orchestra.run.vm04.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-21T07:02:38.700 INFO:teuthology.orchestra.run.vm02.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-21T07:02:38.703 INFO:teuthology.orchestra.run.vm04.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-21T07:02:38.712 INFO:teuthology.orchestra.run.vm02.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-21T07:02:38.715 INFO:teuthology.orchestra.run.vm04.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-21T07:02:38.724 INFO:teuthology.orchestra.run.vm02.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-21T07:02:38.737 INFO:teuthology.orchestra.run.vm07.stdout:0 upgraded, 0 newly installed, 60 to remove and 36 not upgraded. 2026-03-21T07:02:38.737 INFO:teuthology.orchestra.run.vm07.stdout:After this operation, 93.5 MB disk space will be freed. 2026-03-21T07:02:38.739 INFO:teuthology.orchestra.run.vm02.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-21T07:02:38.792 INFO:teuthology.orchestra.run.vm07.stdout:(Reading database ... (Reading database ... 5% (Reading database ... 10% (Reading database ... 15% (Reading database ... 20% (Reading database ... 25% (Reading database ... 30% (Reading database ... 35% (Reading database ... 40% (Reading database ... 45% (Reading database ... 50% (Reading database ... 55% (Reading database ... 60% (Reading database ... 65% (Reading database ... 70% (Reading database ... 75% (Reading database ... 80% (Reading database ... 85% (Reading database ... 90% (Reading database ... 95% (Reading database ... 100% (Reading database ... 123743 files and directories currently installed.) 2026-03-21T07:02:38.795 INFO:teuthology.orchestra.run.vm07.stdout:Removing ceph-mgr-modules-core (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:38.810 INFO:teuthology.orchestra.run.vm07.stdout:Removing jq (1.6-2.1ubuntu3.1) ... 2026-03-21T07:02:38.822 INFO:teuthology.orchestra.run.vm07.stdout:Removing libboost-iostreams1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-21T07:02:38.836 INFO:teuthology.orchestra.run.vm07.stdout:Removing libboost-thread1.74.0:amd64 (1.74.0-14ubuntu3) ... 2026-03-21T07:02:38.848 INFO:teuthology.orchestra.run.vm07.stdout:Removing libcephfs-proxy2 (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:38.859 INFO:teuthology.orchestra.run.vm07.stdout:Removing libthrift-0.16.0:amd64 (0.16.0-2) ... 2026-03-21T07:02:38.871 INFO:teuthology.orchestra.run.vm07.stdout:Removing libqt5network5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T07:02:38.882 INFO:teuthology.orchestra.run.vm07.stdout:Removing libqt5dbus5:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T07:02:38.896 INFO:teuthology.orchestra.run.vm07.stdout:Removing libqt5core5a:amd64 (5.15.3+dfsg-2ubuntu0.2) ... 2026-03-21T07:02:38.922 INFO:teuthology.orchestra.run.vm07.stdout:Removing libdouble-conversion3:amd64 (3.1.7-4) ... 2026-03-21T07:02:38.935 INFO:teuthology.orchestra.run.vm07.stdout:Removing libfuse2:amd64 (2.9.9-5ubuntu3) ... 2026-03-21T07:02:38.947 INFO:teuthology.orchestra.run.vm07.stdout:Removing libgfapi0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.957 INFO:teuthology.orchestra.run.vm07.stdout:Removing libgfrpc0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.969 INFO:teuthology.orchestra.run.vm07.stdout:Removing libgfxdr0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.980 INFO:teuthology.orchestra.run.vm07.stdout:Removing libglusterfs0:amd64 (10.1-1ubuntu0.2) ... 2026-03-21T07:02:38.991 INFO:teuthology.orchestra.run.vm07.stdout:Removing libiscsi7:amd64 (1.19.0-3build2) ... 2026-03-21T07:02:39.002 INFO:teuthology.orchestra.run.vm07.stdout:Removing libjq1:amd64 (1.6-2.1ubuntu3.1) ... 2026-03-21T07:02:39.010 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.011 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.011 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.011 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:38 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.013 INFO:teuthology.orchestra.run.vm07.stdout:Removing liblttng-ust1:amd64 (2.13.1-1ubuntu1) ... 2026-03-21T07:02:39.025 INFO:teuthology.orchestra.run.vm07.stdout:Removing libnbd0 (1.10.5-1) ... 2026-03-21T07:02:39.037 INFO:teuthology.orchestra.run.vm07.stdout:Removing liboath0:amd64 (2.6.7-3ubuntu0.1) ... 2026-03-21T07:02:39.048 INFO:teuthology.orchestra.run.vm07.stdout:Removing libonig5:amd64 (6.9.7.1-2build1) ... 2026-03-21T07:02:39.059 INFO:teuthology.orchestra.run.vm07.stdout:Removing libpcre2-16-0:amd64 (10.39-3ubuntu0.1) ... 2026-03-21T07:02:39.071 INFO:teuthology.orchestra.run.vm07.stdout:Removing libpmemobj1:amd64 (1.11.1-3build1) ... 2026-03-21T07:02:39.082 INFO:teuthology.orchestra.run.vm07.stdout:Removing librdkafka1:amd64 (1.8.0-1build1) ... 2026-03-21T07:02:39.093 INFO:teuthology.orchestra.run.vm07.stdout:Removing nvme-cli (1.16-3ubuntu0.3) ... 2026-03-21T07:02:39.100 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:38 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.100 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:38 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.100 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:38 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.100 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:38 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.100 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:38 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.139 INFO:teuthology.orchestra.run.vm04.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-21T07:02:39.153 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-21T07:02:39.189 INFO:teuthology.orchestra.run.vm02.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-21T07:02:39.204 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-21T07:02:39.215 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-21T07:02:39.269 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-21T07:02:39.387 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:39 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.387 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:39 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.387 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:39 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.387 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:39 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.387 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:39 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.403 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:39 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.403 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:39 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.403 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:39 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.403 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:39 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.404 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:39 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.499 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-21T07:02:39.510 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.510 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.510 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.510 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:39 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.532 INFO:teuthology.orchestra.run.vm07.stdout:Removing python-asyncssh-doc (2.5.0-1ubuntu0.1) ... 2026-03-21T07:02:39.545 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-21T07:02:39.545 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-asyncssh (2.5.0-1ubuntu0.1) ... 2026-03-21T07:02:39.553 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-21T07:02:39.600 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-21T07:02:39.601 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-kubernetes (12.0.1-1ubuntu1) ... 2026-03-21T07:02:39.601 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:39.650 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:39.651 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:39.701 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:39.707 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-21T07:02:39.756 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-21T07:02:39.770 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-21T07:02:39.818 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-21T07:02:39.823 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-21T07:02:39.862 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-google-auth (1.5.1-3) ... 2026-03-21T07:02:39.874 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-21T07:02:39.877 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-21T07:02:39.902 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:39 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.903 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:39 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.903 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:39 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:39 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.903 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:39 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:39.922 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-cachetools (5.0.0-1) ... 2026-03-21T07:02:39.928 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-21T07:02:39.943 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-21T07:02:39.978 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-ceph-argparse (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:39.996 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-21T07:02:39.998 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-21T07:02:40.032 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-ceph-common (20.2.0-712-g70f8415b-1jammy) ... 2026-03-21T07:02:40.048 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-21T07:02:40.052 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-21T07:02:40.089 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-cherrypy3 (18.6.1-4) ... 2026-03-21T07:02:40.103 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-21T07:02:40.103 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-21T07:02:40.156 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-cheroot (8.5.2+ds1-1ubuntu3.1) ... 2026-03-21T07:02:40.156 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-21T07:02:40.159 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-21T07:02:40.209 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-21T07:02:40.211 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-jaraco.collections (3.4.0-2) ... 2026-03-21T07:02:40.261 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-jaraco.classes (3.2.1-3) ... 2026-03-21T07:02:40.279 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-21T07:02:40.320 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-portend (3.0.0-1) ... 2026-03-21T07:02:40.332 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-21T07:02:40.342 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-21T07:02:40.368 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-tempora (4.1.2-1) ... 2026-03-21T07:02:40.394 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-21T07:02:40.397 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-21T07:02:40.417 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-jaraco.text (3.6.0-2) ... 2026-03-21T07:02:40.450 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-21T07:02:40.450 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-21T07:02:40.463 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-jaraco.functools (3.4.0-2) ... 2026-03-21T07:02:40.503 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-21T07:02:40.508 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-21T07:02:40.510 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-sklearn (0.23.2-5ubuntu6) ... 2026-03-21T07:02:40.561 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-21T07:02:40.563 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-21T07:02:40.618 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-21T07:02:40.618 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-21T07:02:40.635 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-joblib (0.17.0-4ubuntu1) ... 2026-03-21T07:02:40.667 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-21T07:02:40.674 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-rsa (4.8-1) ... 2026-03-21T07:02:40.698 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-natsort (8.0.2-1) ... 2026-03-21T07:02:40.719 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-rsa (4.8-1) ... 2026-03-21T07:02:40.729 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-21T07:02:40.749 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-prettytable (2.5.0-2) ... 2026-03-21T07:02:40.766 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-21T07:02:40.786 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-21T07:02:40.801 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-21T07:02:40.804 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-psutil (5.9.0-1build1) ... 2026-03-21T07:02:40.826 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-21T07:02:40.841 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-21T07:02:40.853 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-21T07:02:40.858 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-routes (2.5.1-1ubuntu1) ... 2026-03-21T07:02:40.893 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-21T07:02:40.906 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-21T07:02:40.913 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-repoze.lru (0.7-2) ... 2026-03-21T07:02:40.947 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-21T07:02:40.959 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-21T07:02:40.964 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-requests-oauthlib (1.3.0+ds-0.1) ... 2026-03-21T07:02:41.002 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-21T07:02:41.017 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-rsa (4.8-1) ... 2026-03-21T07:02:41.017 INFO:teuthology.orchestra.run.vm04.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-21T07:02:41.062 INFO:teuthology.orchestra.run.vm02.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-21T07:02:41.068 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-simplejson (3.17.6-1build1) ... 2026-03-21T07:02:41.071 INFO:teuthology.orchestra.run.vm04.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-21T07:02:41.093 INFO:teuthology.orchestra.run.vm04.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-21T07:02:41.116 INFO:teuthology.orchestra.run.vm02.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-21T07:02:41.125 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-sklearn-lib:amd64 (0.23.2-5ubuntu6) ... 2026-03-21T07:02:41.138 INFO:teuthology.orchestra.run.vm02.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-21T07:02:41.141 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-threadpoolctl (3.1.0-1) ... 2026-03-21T07:02:41.193 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-wcwidth (0.2.5+dfsg1-1) ... 2026-03-21T07:02:41.245 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-webob (1:1.8.6-1.1ubuntu0.1) ... 2026-03-21T07:02:41.300 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-websocket (1.2.3-1) ... 2026-03-21T07:02:41.357 INFO:teuthology.orchestra.run.vm07.stdout:Removing python3-zc.lockfile (2.0-1) ... 2026-03-21T07:02:41.409 INFO:teuthology.orchestra.run.vm07.stdout:Removing qttranslations5-l10n (5.15.3-1) ... 2026-03-21T07:02:41.431 INFO:teuthology.orchestra.run.vm07.stdout:Removing smartmontools (7.2-1ubuntu0.1) ... 2026-03-21T07:02:41.440 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.440 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.441 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.441 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.493 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:41 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.493 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:41 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.493 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:41 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.493 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:41 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.493 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:41 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.521 INFO:teuthology.orchestra.run.vm04.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-21T07:02:41.533 INFO:teuthology.orchestra.run.vm04.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-21T07:02:41.564 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:41.577 INFO:teuthology.orchestra.run.vm02.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-21T07:02:41.589 INFO:teuthology.orchestra.run.vm02.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-21T07:02:41.627 INFO:teuthology.orchestra.run.vm04.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:41.638 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:41.710 INFO:teuthology.orchestra.run.vm02.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:41.760 INFO:journalctl@ceph.mon.b.vm04.stdout:Mar 21 07:02:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.760 INFO:journalctl@ceph.osd.2.vm04.stdout:Mar 21 07:02:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.760 INFO:journalctl@ceph.osd.3.vm04.stdout:Mar 21 07:02:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.760 INFO:journalctl@ceph.osd.4.vm04.stdout:Mar 21 07:02:41 vm04 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.763 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.763 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.763 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.763 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.763 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.856 INFO:teuthology.orchestra.run.vm07.stdout:Removing socat (1.7.4.1-3ubuntu4) ... 2026-03-21T07:02:41.869 INFO:teuthology.orchestra.run.vm07.stdout:Removing xmlstarlet (1.6.1-2.1) ... 2026-03-21T07:02:41.887 INFO:journalctl@ceph.mgr.x.vm02.stdout:Mar 21 07:02:41 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.887 INFO:journalctl@ceph.osd.0.vm02.stdout:Mar 21 07:02:41 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.887 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:41 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.887 INFO:journalctl@ceph.mon.a.vm02.stdout:Mar 21 07:02:41 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.887 INFO:journalctl@ceph.osd.1.vm02.stdout:Mar 21 07:02:41 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:41.901 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for man-db (2.10.2-1) ... 2026-03-21T07:02:41.946 INFO:teuthology.orchestra.run.vm07.stdout:Processing triggers for libc-bin (2.35-0ubuntu3.13) ... 2026-03-21T07:02:42.153 INFO:journalctl@ceph.mon.c.vm07.stdout:Mar 21 07:02:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:42.153 INFO:journalctl@ceph.osd.5.vm07.stdout:Mar 21 07:02:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:42.153 INFO:journalctl@ceph.osd.6.vm07.stdout:Mar 21 07:02:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:42.153 INFO:journalctl@ceph.osd.7.vm07.stdout:Mar 21 07:02:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:42.153 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:41 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:43.004 INFO:teuthology.orchestra.run.vm04.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:43.007 DEBUG:teuthology.parallel:result is None 2026-03-21T07:02:43.134 INFO:teuthology.orchestra.run.vm02.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:43.138 DEBUG:teuthology.parallel:result is None 2026-03-21T07:02:43.218 INFO:teuthology.orchestra.run.vm07.stderr:W: --force-yes is deprecated, use one of the options starting with --allow instead. 2026-03-21T07:02:43.221 DEBUG:teuthology.parallel:result is None 2026-03-21T07:02:43.221 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm02.local 2026-03-21T07:02:43.221 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm04.local 2026-03-21T07:02:43.221 INFO:teuthology.task.install:Removing ceph sources lists on ubuntu@vm07.local 2026-03-21T07:02:43.221 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-21T07:02:43.222 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-21T07:02:43.222 DEBUG:teuthology.orchestra.run.vm07:> sudo rm -f /etc/apt/sources.list.d/ceph.list 2026-03-21T07:02:43.229 DEBUG:teuthology.orchestra.run.vm04:> sudo apt-get update 2026-03-21T07:02:43.232 DEBUG:teuthology.orchestra.run.vm02:> sudo apt-get update 2026-03-21T07:02:43.272 DEBUG:teuthology.orchestra.run.vm07:> sudo apt-get update 2026-03-21T07:02:43.336 INFO:teuthology.orchestra.run.vm04.stdout:Hit:1 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-21T07:02:43.346 INFO:teuthology.orchestra.run.vm02.stdout:Hit:1 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-21T07:02:43.391 INFO:teuthology.orchestra.run.vm07.stdout:Hit:1 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-21T07:02:43.392 INFO:teuthology.orchestra.run.vm07.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-21T07:02:43.400 INFO:teuthology.orchestra.run.vm07.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-21T07:02:43.519 INFO:teuthology.orchestra.run.vm04.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-21T07:02:43.531 INFO:teuthology.orchestra.run.vm02.stdout:Hit:2 http://archive.ubuntu.com/ubuntu jammy InRelease 2026-03-21T07:02:43.568 INFO:teuthology.orchestra.run.vm07.stdout:Hit:4 http://security.ubuntu.com/ubuntu jammy-security InRelease 2026-03-21T07:02:43.616 INFO:teuthology.orchestra.run.vm04.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-21T07:02:43.632 INFO:teuthology.orchestra.run.vm02.stdout:Hit:3 http://archive.ubuntu.com/ubuntu jammy-updates InRelease 2026-03-21T07:02:43.713 INFO:teuthology.orchestra.run.vm04.stdout:Hit:4 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-21T07:02:43.733 INFO:teuthology.orchestra.run.vm02.stdout:Hit:4 http://archive.ubuntu.com/ubuntu jammy-backports InRelease 2026-03-21T07:02:44.304 INFO:teuthology.orchestra.run.vm07.stdout:Reading package lists... 2026-03-21T07:02:44.316 DEBUG:teuthology.parallel:result is None 2026-03-21T07:02:44.476 INFO:teuthology.orchestra.run.vm04.stdout:Reading package lists... 2026-03-21T07:02:44.490 INFO:teuthology.orchestra.run.vm02.stdout:Reading package lists... 2026-03-21T07:02:44.491 DEBUG:teuthology.parallel:result is None 2026-03-21T07:02:44.502 DEBUG:teuthology.parallel:result is None 2026-03-21T07:02:44.502 DEBUG:teuthology.run_tasks:Unwinding manager cephadm 2026-03-21T07:02:44.504 INFO:tasks.cephadm:Teardown begin 2026-03-21T07:02:44.504 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-21T07:02:44.511 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-21T07:02:44.518 DEBUG:teuthology.orchestra.run.vm07:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-21T07:02:44.526 INFO:tasks.cephadm:Disabling cephadm mgr module 2026-03-21T07:02:44.526 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm --image quay.ceph.io/ceph-ci/ceph:70f8415b300f041766fa27faf7d5472699e32388 shell -c /etc/ceph/ceph.conf -k /etc/ceph/ceph.client.admin.keyring --fsid b16ecafc-24f1-11f1-8ede-8330751617ee -- ceph mgr module disable cephadm 2026-03-21T07:02:44.796 INFO:teuthology.orchestra.run.vm02.stderr:Inferring config /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/mon.a/config 2026-03-21T07:02:45.178 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-21T07:02:45.176+0000 7fdf8831c640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-21T07:02:45.178 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-21T07:02:45.176+0000 7fdf8831c640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-21T07:02:45.178 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-21T07:02:45.176+0000 7fdf8831c640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-21T07:02:45.178 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-21T07:02:45.176+0000 7fdf8831c640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-21T07:02:45.178 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-21T07:02:45.176+0000 7fdf8831c640 -1 auth: error reading file: /etc/ceph/ceph.keyring: bufferlist::read_file(/etc/ceph/ceph.keyring): read error:(21) Is a directory 2026-03-21T07:02:45.178 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-21T07:02:45.176+0000 7fdf8831c640 -1 auth: failed to load /etc/ceph/ceph.keyring: (21) Is a directory 2026-03-21T07:02:45.178 INFO:teuthology.orchestra.run.vm02.stderr:2026-03-21T07:02:45.176+0000 7fdf8831c640 -1 monclient: keyring not found 2026-03-21T07:02:45.178 INFO:teuthology.orchestra.run.vm02.stderr:[errno 21] error connecting to the cluster 2026-03-21T07:02:45.228 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-21T07:02:45.228 INFO:tasks.cephadm:Cleaning up testdir ceph.* files... 2026-03-21T07:02:45.228 DEBUG:teuthology.orchestra.run.vm02:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-21T07:02:45.232 DEBUG:teuthology.orchestra.run.vm04:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-21T07:02:45.236 DEBUG:teuthology.orchestra.run.vm07:> rm -f /home/ubuntu/cephtest/seed.ceph.conf /home/ubuntu/cephtest/ceph.pub 2026-03-21T07:02:45.239 INFO:tasks.cephadm:Stopping all daemons... 2026-03-21T07:02:45.240 INFO:tasks.cephadm.mon.a:Stopping mon.a... 2026-03-21T07:02:45.240 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.a 2026-03-21T07:02:45.284 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.a.service' 2026-03-21T07:02:45.339 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:45.339 INFO:tasks.cephadm.mon.a:Stopped mon.a 2026-03-21T07:02:45.339 INFO:tasks.cephadm.mon.c:Stopping mon.b... 2026-03-21T07:02:45.339 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.b 2026-03-21T07:02:45.350 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.b.service' 2026-03-21T07:02:45.404 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:45.405 INFO:tasks.cephadm.mon.c:Stopped mon.b 2026-03-21T07:02:45.405 INFO:tasks.cephadm.mon.c:Stopping mon.c... 2026-03-21T07:02:45.405 DEBUG:teuthology.orchestra.run.vm07:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.c 2026-03-21T07:02:45.415 DEBUG:teuthology.orchestra.run.vm07:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mon.c.service' 2026-03-21T07:02:45.469 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:45.469 INFO:tasks.cephadm.mon.c:Stopped mon.c 2026-03-21T07:02:45.469 INFO:tasks.cephadm.mgr.x:Stopping mgr.x... 2026-03-21T07:02:45.469 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mgr.x 2026-03-21T07:02:45.478 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@mgr.x.service' 2026-03-21T07:02:45.529 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:45.529 INFO:tasks.cephadm.mgr.x:Stopped mgr.x 2026-03-21T07:02:45.530 INFO:tasks.cephadm.osd.0:Stopping osd.0... 2026-03-21T07:02:45.530 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.0 2026-03-21T07:02:45.580 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.0.service' 2026-03-21T07:02:45.635 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:45.635 INFO:tasks.cephadm.osd.0:Stopped osd.0 2026-03-21T07:02:45.635 INFO:tasks.cephadm.osd.1:Stopping osd.1... 2026-03-21T07:02:45.635 DEBUG:teuthology.orchestra.run.vm02:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.1 2026-03-21T07:02:45.688 DEBUG:teuthology.orchestra.run.vm02:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.1.service' 2026-03-21T07:02:45.742 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:45.742 INFO:tasks.cephadm.osd.1:Stopped osd.1 2026-03-21T07:02:45.742 INFO:tasks.cephadm.osd.2:Stopping osd.2... 2026-03-21T07:02:45.742 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.2 2026-03-21T07:02:45.751 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.2.service' 2026-03-21T07:02:45.807 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:45.807 INFO:tasks.cephadm.osd.2:Stopped osd.2 2026-03-21T07:02:45.807 INFO:tasks.cephadm.osd.3:Stopping osd.3... 2026-03-21T07:02:45.807 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.3 2026-03-21T07:02:45.860 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.3.service' 2026-03-21T07:02:45.914 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:45.914 INFO:tasks.cephadm.osd.3:Stopped osd.3 2026-03-21T07:02:45.914 INFO:tasks.cephadm.osd.4:Stopping osd.4... 2026-03-21T07:02:45.914 DEBUG:teuthology.orchestra.run.vm04:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.4 2026-03-21T07:02:45.964 DEBUG:teuthology.orchestra.run.vm04:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.4.service' 2026-03-21T07:02:46.032 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:46.032 INFO:tasks.cephadm.osd.4:Stopped osd.4 2026-03-21T07:02:46.032 INFO:tasks.cephadm.osd.5:Stopping osd.5... 2026-03-21T07:02:46.033 DEBUG:teuthology.orchestra.run.vm07:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.5 2026-03-21T07:02:46.044 DEBUG:teuthology.orchestra.run.vm07:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.5.service' 2026-03-21T07:02:46.100 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:46.101 INFO:tasks.cephadm.osd.5:Stopped osd.5 2026-03-21T07:02:46.101 INFO:tasks.cephadm.osd.6:Stopping osd.6... 2026-03-21T07:02:46.101 DEBUG:teuthology.orchestra.run.vm07:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.6 2026-03-21T07:02:46.154 DEBUG:teuthology.orchestra.run.vm07:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.6.service' 2026-03-21T07:02:46.208 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:46.208 INFO:tasks.cephadm.osd.6:Stopped osd.6 2026-03-21T07:02:46.208 INFO:tasks.cephadm.osd.7:Stopping osd.7... 2026-03-21T07:02:46.208 DEBUG:teuthology.orchestra.run.vm07:> sudo systemctl stop ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.7 2026-03-21T07:02:46.258 DEBUG:teuthology.orchestra.run.vm07:> sudo pkill -f 'journalctl -f -n 0 -u ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@osd.7.service' 2026-03-21T07:02:46.312 DEBUG:teuthology.orchestra.run:got remote process result: None 2026-03-21T07:02:46.312 INFO:tasks.cephadm.osd.7:Stopped osd.7 2026-03-21T07:02:46.312 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid b16ecafc-24f1-11f1-8ede-8330751617ee --force --keep-logs 2026-03-21T07:02:46.582 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T07:02:46.887 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:46 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:47.183 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:46 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:47.183 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:47 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:47.529 INFO:journalctl@ceph.iscsi.iscsi.a.vm02.stdout:Mar 21 07:02:47 vm02 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:48.270 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid b16ecafc-24f1-11f1-8ede-8330751617ee --force --keep-logs 2026-03-21T07:02:48.531 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T07:02:49.761 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid b16ecafc-24f1-11f1-8ede-8330751617ee --force --keep-logs 2026-03-21T07:02:50.018 INFO:teuthology.orchestra.run.vm07.stdout:Deleting cluster with fsid: b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T07:02:50.337 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:50 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:50.589 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:50 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:50.589 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:50 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:50.903 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:50 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:51.240 INFO:journalctl@ceph.iscsi.iscsi.b.vm07.stdout:Mar 21 07:02:50 vm07 systemd[1]: /etc/systemd/system/ceph-b16ecafc-24f1-11f1-8ede-8330751617ee@.service:23: Unit configured to use KillMode=none. This is unsafe, as it disables systemd's process lifecycle management for the service. Please update your service to use a safer KillMode=, such as 'mixed' or 'control-group'. Support for KillMode=none is deprecated and will eventually be removed. 2026-03-21T07:02:51.647 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-21T07:02:51.655 INFO:teuthology.orchestra.run.vm02.stderr:rm: cannot remove '/etc/ceph/ceph.client.admin.keyring': Is a directory 2026-03-21T07:02:51.655 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-21T07:02:51.656 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-21T07:02:51.664 DEBUG:teuthology.orchestra.run.vm07:> sudo rm -f /etc/ceph/ceph.conf /etc/ceph/ceph.client.admin.keyring 2026-03-21T07:02:51.673 INFO:tasks.cephadm:Archiving crash dumps... 2026-03-21T07:02:51.673 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/crash to /archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458/remote/vm02/crash 2026-03-21T07:02:51.673 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/crash -- . 2026-03-21T07:02:51.708 INFO:teuthology.orchestra.run.vm02.stderr:tar: /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/crash: Cannot open: No such file or directory 2026-03-21T07:02:51.708 INFO:teuthology.orchestra.run.vm02.stderr:tar: Error is not recoverable: exiting now 2026-03-21T07:02:51.709 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/crash to /archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458/remote/vm04/crash 2026-03-21T07:02:51.709 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/crash -- . 2026-03-21T07:02:51.717 INFO:teuthology.orchestra.run.vm04.stderr:tar: /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/crash: Cannot open: No such file or directory 2026-03-21T07:02:51.717 INFO:teuthology.orchestra.run.vm04.stderr:tar: Error is not recoverable: exiting now 2026-03-21T07:02:51.718 DEBUG:teuthology.misc:Transferring archived files from vm07:/var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/crash to /archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458/remote/vm07/crash 2026-03-21T07:02:51.718 DEBUG:teuthology.orchestra.run.vm07:> sudo tar c -f - -C /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/crash -- . 2026-03-21T07:02:51.726 INFO:teuthology.orchestra.run.vm07.stderr:tar: /var/lib/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/crash: Cannot open: No such file or directory 2026-03-21T07:02:51.726 INFO:teuthology.orchestra.run.vm07.stderr:tar: Error is not recoverable: exiting now 2026-03-21T07:02:51.727 INFO:tasks.cephadm:Checking cluster log for badness... 2026-03-21T07:02:51.727 DEBUG:teuthology.orchestra.run.vm02:> sudo egrep '\[ERR\]|\[WRN\]|\[SEC\]' /var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph.log | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | head -n 1 2026-03-21T07:02:51.759 INFO:teuthology.orchestra.run.vm02.stderr:grep: /var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph.log: No such file or directory 2026-03-21T07:02:51.760 WARNING:tasks.cephadm:Found errors (ERR|WRN|SEC) in cluster log 2026-03-21T07:02:51.760 DEBUG:teuthology.orchestra.run.vm02:> sudo egrep '\[SEC\]' /var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph.log | egrep -v '\(MDS_ALL_DOWN\)' | egrep -v '\(MDS_UP_LESS_THAN_MAX\)' | egrep -v MON_DOWN | head -n 1 2026-03-21T07:02:51.811 INFO:teuthology.orchestra.run.vm02.stderr:grep: /var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph.log: No such file or directory 2026-03-21T07:02:51.813 INFO:tasks.cephadm:Compressing logs... 2026-03-21T07:02:51.813 DEBUG:teuthology.orchestra.run.vm02:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-21T07:02:51.857 DEBUG:teuthology.orchestra.run.vm04:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-21T07:02:51.859 DEBUG:teuthology.orchestra.run.vm07:> time sudo find /var/log/ceph /var/log/rbd-target-api -name '*.log' -print0 | sudo xargs --max-args=1 --max-procs=0 --verbose -0 --no-run-if-empty -- gzip -5 --verbose -- 2026-03-21T07:02:51.869 INFO:teuthology.orchestra.run.vm02.stderr:find: ‘/var/log/rbd-target-api’: No such file or directory 2026-03-21T07:02:51.870 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph-volume.log 2026-03-21T07:02:51.871 INFO:teuthology.orchestra.run.vm04.stderr:find: gzip -5 --verbose -- /var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph-volume.log 2026-03-21T07:02:51.871 INFO:teuthology.orchestra.run.vm04.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-21T07:02:51.871 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-21T07:02:51.872 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-21T07:02:51.872 INFO:teuthology.orchestra.run.vm04.stderr:/var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph-volume.log: /var/log/ceph/cephadm.log: 79.6% -- replaced with /var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph-volume.log.gz 2026-03-21T07:02:51.872 INFO:teuthology.orchestra.run.vm04.stderr: 89.6% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-21T07:02:51.873 INFO:teuthology.orchestra.run.vm02.stderr:/var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph-volume.log: /var/log/ceph/cephadm.log: 73.4% -- replaced with /var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph-volume.log.gz 2026-03-21T07:02:51.873 INFO:teuthology.orchestra.run.vm02.stderr: 87.5% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-21T07:02:51.874 INFO:teuthology.orchestra.run.vm04.stderr: 2026-03-21T07:02:51.874 INFO:teuthology.orchestra.run.vm04.stderr:real 0m0.011s 2026-03-21T07:02:51.874 INFO:teuthology.orchestra.run.vm04.stderr:user 0m0.003s 2026-03-21T07:02:51.874 INFO:teuthology.orchestra.run.vm04.stderr:sys 0m0.017s 2026-03-21T07:02:51.874 INFO:teuthology.orchestra.run.vm07.stderr:find: gzip -5 --verbose -- /var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph-volume.log 2026-03-21T07:02:51.874 INFO:teuthology.orchestra.run.vm07.stderr:‘/var/log/rbd-target-api’: No such file or directory 2026-03-21T07:02:51.875 INFO:teuthology.orchestra.run.vm02.stderr: 2026-03-21T07:02:51.875 INFO:teuthology.orchestra.run.vm02.stderr:real 0m0.015s 2026-03-21T07:02:51.875 INFO:teuthology.orchestra.run.vm02.stderr:user 0m0.010s 2026-03-21T07:02:51.875 INFO:teuthology.orchestra.run.vm02.stderr:sys 0m0.011s 2026-03-21T07:02:51.875 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /var/log/ceph/cephadm.log 2026-03-21T07:02:51.875 INFO:teuthology.orchestra.run.vm07.stderr:/var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph-volume.log: 79.3%/var/log/ceph/cephadm.log: -- replaced with /var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph-volume.log.gz 2026-03-21T07:02:51.876 INFO:teuthology.orchestra.run.vm07.stderr: 90.5% -- replaced with /var/log/ceph/cephadm.log.gz 2026-03-21T07:02:51.877 INFO:teuthology.orchestra.run.vm07.stderr: 2026-03-21T07:02:51.877 INFO:teuthology.orchestra.run.vm07.stderr:real 0m0.012s 2026-03-21T07:02:51.877 INFO:teuthology.orchestra.run.vm07.stderr:user 0m0.009s 2026-03-21T07:02:51.877 INFO:teuthology.orchestra.run.vm07.stderr:sys 0m0.011s 2026-03-21T07:02:51.877 INFO:tasks.cephadm:Archiving logs... 2026-03-21T07:02:51.877 DEBUG:teuthology.misc:Transferring archived files from vm02:/var/log/ceph to /archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458/remote/vm02/log 2026-03-21T07:02:51.877 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-21T07:02:51.926 DEBUG:teuthology.misc:Transferring archived files from vm04:/var/log/ceph to /archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458/remote/vm04/log 2026-03-21T07:02:51.926 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-21T07:02:51.935 DEBUG:teuthology.misc:Transferring archived files from vm07:/var/log/ceph to /archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458/remote/vm07/log 2026-03-21T07:02:51.935 DEBUG:teuthology.orchestra.run.vm07:> sudo tar c -f - -C /var/log/ceph -- . 2026-03-21T07:02:51.944 INFO:tasks.cephadm:Removing cluster... 2026-03-21T07:02:51.944 DEBUG:teuthology.orchestra.run.vm02:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid b16ecafc-24f1-11f1-8ede-8330751617ee --force 2026-03-21T07:02:52.221 INFO:teuthology.orchestra.run.vm02.stdout:Deleting cluster with fsid: b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T07:02:52.265 DEBUG:teuthology.orchestra.run.vm04:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid b16ecafc-24f1-11f1-8ede-8330751617ee --force 2026-03-21T07:02:52.501 INFO:teuthology.orchestra.run.vm04.stdout:Deleting cluster with fsid: b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T07:02:52.545 DEBUG:teuthology.orchestra.run.vm07:> sudo /home/ubuntu/cephtest/cephadm rm-cluster --fsid b16ecafc-24f1-11f1-8ede-8330751617ee --force 2026-03-21T07:02:52.784 INFO:teuthology.orchestra.run.vm07.stdout:Deleting cluster with fsid: b16ecafc-24f1-11f1-8ede-8330751617ee 2026-03-21T07:02:52.830 INFO:tasks.cephadm:Removing cephadm ... 2026-03-21T07:02:52.830 DEBUG:teuthology.orchestra.run.vm02:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-21T07:02:52.834 DEBUG:teuthology.orchestra.run.vm04:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-21T07:02:52.838 DEBUG:teuthology.orchestra.run.vm07:> rm -rf /home/ubuntu/cephtest/cephadm 2026-03-21T07:02:52.842 INFO:tasks.cephadm:Teardown complete 2026-03-21T07:02:52.842 DEBUG:teuthology.run_tasks:Unwinding manager clock 2026-03-21T07:02:52.844 INFO:teuthology.task.clock:Checking final clock skew... 2026-03-21T07:02:52.844 DEBUG:teuthology.orchestra.run.vm02:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-21T07:02:52.876 DEBUG:teuthology.orchestra.run.vm04:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-21T07:02:52.880 DEBUG:teuthology.orchestra.run.vm07:> PATH=/usr/bin:/usr/sbin ntpq -p || PATH=/usr/bin:/usr/sbin chronyc sources || true 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout: remote refid st t when poll reach delay offset jitter 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:============================================================================== 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:-static.179.181. 161.62.157.173 3 u 31 256 377 23.501 +0.378 0.814 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:+ntp2.uni-ulm.de 129.69.253.1 2 u 26 128 377 28.055 -0.541 0.784 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:+time.cloudflare 10.216.8.4 3 u 40 64 377 20.371 +1.487 0.727 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:-91.198.32.34 (9 .PPS. 1 u 27 128 377 25.270 +1.864 1.626 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:-185.232.69.65 ( .PHC0. 1 u 29 128 377 28.290 -3.003 1.123 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:*srv02.spectre-n 131.188.3.222 2 u 36 64 377 23.860 +0.989 0.952 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:#64.188.67.160 195.176.26.215 2 u 31 128 377 26.636 -2.990 1.206 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:-185.125.190.57 29.88.99.4 2 u 59 64 377 31.270 -0.171 1.149 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:-vps-nue1.orlean 195.145.119.188 2 u 38 64 377 28.732 -1.879 0.762 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:-185.125.190.58 145.238.80.80 2 u 61 64 377 35.100 -0.982 0.674 2026-03-21T07:02:53.148 INFO:teuthology.orchestra.run.vm07.stdout:-185.125.190.56 194.121.207.249 2 u 125 64 376 35.354 -0.859 0.610 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout: remote refid st t when poll reach delay offset jitter 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:============================================================================== 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:*ntp2.uni-ulm.de 129.69.253.1 2 u 40 64 377 27.338 +1.245 0.544 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:+64.188.67.160 195.176.26.215 2 u 32 64 377 26.495 -0.012 0.517 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:+vps-nue1.orlean 195.145.119.188 2 u 31 64 377 28.200 -0.692 0.931 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:#139-162-156-95. 82.35.162.146 2 u 20 64 377 22.632 -1.468 1.258 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:-91.198.32.34 (9 .PPS. 1 u 36 64 377 25.358 +4.476 0.507 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:#185.232.69.65 ( .PHC0. 1 u 34 64 377 28.229 -1.063 0.979 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:-static.179.181. 161.62.157.173 3 u 35 64 377 23.517 +1.829 0.884 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:+srv02.spectre-n 131.188.3.222 2 u 32 64 377 23.833 +1.517 0.996 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:#ns.gunnarhofman 124.216.164.14 2 u 28 64 377 24.957 +1.403 0.919 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:#alphyn.canonica 132.163.96.1 2 u 65 64 377 115.935 -6.914 0.996 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:#157.90.247.99 ( 158.75.5.245 2 u 29 64 377 25.235 +4.565 0.740 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:#formularfetisch 131.188.3.222 2 u 34 64 377 25.038 +1.830 0.596 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:+185.125.190.57 29.88.99.4 2 u 60 64 377 35.273 +0.213 0.551 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:-time.cloudflare 10.124.8.190 3 u 34 64 377 20.360 +2.590 0.968 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:+185.125.190.56 194.121.207.249 2 u 54 64 377 35.096 +0.992 0.537 2026-03-21T07:02:53.149 INFO:teuthology.orchestra.run.vm04.stdout:+185.125.190.58 145.238.80.80 2 u 57 64 377 35.364 -0.110 0.937 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout: remote refid st t when poll reach delay offset jitter 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:============================================================================== 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout: 0.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout: 1.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout: 2.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout: 3.ubuntu.pool.n .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout: ntp.ubuntu.com .POOL. 16 p - 64 0 0.000 +0.000 0.000 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:+ntp2.uni-ulm.de 129.69.253.1 2 u 24 64 377 27.436 -4.815 1.645 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:#static.179.181. 161.62.157.173 3 u 32 64 377 23.539 -2.709 1.612 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:#srv02.spectre-n 131.188.3.222 2 u 35 64 377 23.859 -4.024 2.027 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:+139-162-156-95. 82.35.162.146 2 u 30 64 377 22.490 -7.987 1.444 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:#formularfetisch 131.188.3.222 2 u 29 64 377 25.023 -4.391 2.043 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:+64.188.67.160 195.176.26.215 2 u 34 64 377 26.454 -5.505 1.357 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:#time.cloudflare 10.125.9.225 3 u 24 64 377 20.416 -0.322 2.162 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:*91.198.32.34 (9 .PPS. 1 u 32 64 377 27.754 -3.194 1.868 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:+185.232.69.65 ( .PHC0. 1 u 27 64 377 28.265 -5.601 1.391 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:#alphyn.canonica 132.163.96.1 2 u 64 64 377 102.020 -6.064 1.609 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:+ns.gunnarhofman 124.216.164.14 2 u 34 64 377 24.914 -4.299 2.014 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:+vps-nue1.orlean 195.145.119.188 2 u 26 64 377 28.665 -2.955 2.927 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:+185.125.190.56 194.121.207.249 2 u 51 64 377 36.396 -3.389 1.466 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:#157.90.247.99 ( 158.75.5.245 2 u 29 64 377 25.244 -2.164 1.938 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:+185.125.190.58 145.238.80.80 2 u 62 64 377 35.137 -5.747 2.210 2026-03-21T07:02:53.150 INFO:teuthology.orchestra.run.vm02.stdout:+185.125.190.57 29.88.99.4 2 u 47 64 377 36.521 -5.094 1.987 2026-03-21T07:02:53.151 DEBUG:teuthology.run_tasks:Unwinding manager ansible.cephlab 2026-03-21T07:02:53.153 INFO:teuthology.task.ansible:Skipping ansible cleanup... 2026-03-21T07:02:53.153 DEBUG:teuthology.run_tasks:Unwinding manager selinux 2026-03-21T07:02:53.156 DEBUG:teuthology.run_tasks:Unwinding manager pcp 2026-03-21T07:02:53.158 DEBUG:teuthology.run_tasks:Unwinding manager internal.timer 2026-03-21T07:02:53.160 INFO:teuthology.task.internal:Duration was 1128.932640 seconds 2026-03-21T07:02:53.160 DEBUG:teuthology.run_tasks:Unwinding manager internal.syslog 2026-03-21T07:02:53.162 INFO:teuthology.task.internal.syslog:Shutting down syslog monitoring... 2026-03-21T07:02:53.162 DEBUG:teuthology.orchestra.run.vm02:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-21T07:02:53.163 DEBUG:teuthology.orchestra.run.vm04:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-21T07:02:53.164 DEBUG:teuthology.orchestra.run.vm07:> sudo rm -f -- /etc/rsyslog.d/80-cephtest.conf && sudo service rsyslog restart 2026-03-21T07:02:53.192 INFO:teuthology.task.internal.syslog:Checking logs for errors... 2026-03-21T07:02:53.192 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm02.local 2026-03-21T07:02:53.192 DEBUG:teuthology.orchestra.run.vm02:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-21T07:02:53.244 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm04.local 2026-03-21T07:02:53.244 DEBUG:teuthology.orchestra.run.vm04:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-21T07:02:53.252 DEBUG:teuthology.task.internal.syslog:Checking ubuntu@vm07.local 2026-03-21T07:02:53.252 DEBUG:teuthology.orchestra.run.vm07:> grep -E --binary-files=text '\bBUG\b|\bINFO\b|\bDEADLOCK\b' /home/ubuntu/cephtest/archive/syslog/kern.log | grep -v 'task .* blocked for more than .* seconds' | grep -v 'lockdep is turned off' | grep -v 'trying to register non-static key' | grep -v 'DEBUG: fsize' | grep -v CRON | grep -v 'BUG: bad unlock balance detected' | grep -v 'inconsistent lock state' | grep -v '*** DEADLOCK ***' | grep -v 'INFO: possible irq lock inversion dependency detected' | grep -v 'INFO: NMI handler (perf_event_nmi_handler) took too long to run' | grep -v 'INFO: recovery required on readonly' | grep -v 'ceph-create-keys: INFO' | grep -v INFO:ceph-create-keys | grep -v 'Loaded datasource DataSourceOpenStack' | grep -v 'container-storage-setup: INFO: Volume group backing root filesystem could not be determined' | grep -E -v '\bsalt-master\b|\bsalt-minion\b|\bsalt-api\b' | grep -v ceph-crash | grep -E -v '\btcmu-runner\b.*\bINFO\b' | head -n 1 2026-03-21T07:02:53.261 INFO:teuthology.task.internal.syslog:Gathering journactl... 2026-03-21T07:02:53.261 DEBUG:teuthology.orchestra.run.vm02:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-21T07:02:53.288 DEBUG:teuthology.orchestra.run.vm04:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-21T07:02:53.295 DEBUG:teuthology.orchestra.run.vm07:> sudo journalctl > /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-21T07:02:53.358 INFO:teuthology.task.internal.syslog:Compressing syslogs... 2026-03-21T07:02:53.358 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-21T07:02:53.360 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-21T07:02:53.361 DEBUG:teuthology.orchestra.run.vm07:> find /home/ubuntu/cephtest/archive/syslog -name '*.log' -print0 | sudo xargs -0 --max-args=1 --max-procs=0 --verbose --no-run-if-empty -- gzip -5 --verbose -- 2026-03-21T07:02:53.367 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-21T07:02:53.368 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-21T07:02:53.368 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-21T07:02:53.368 INFO:teuthology.orchestra.run.vm02.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-21T07:02:53.368 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-21T07:02:53.368 INFO:teuthology.orchestra.run.vm04.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-21T07:02:53.369 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-21T07:02:53.369 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: gzip -5 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-21T07:02:53.369 INFO:teuthology.orchestra.run.vm02.stderr: --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-21T07:02:53.369 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-21T07:02:53.369 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/kern.log 2026-03-21T07:02:53.370 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/misc.log 2026-03-21T07:02:53.370 INFO:teuthology.orchestra.run.vm07.stderr:/home/ubuntu/cephtest/archive/syslog/kern.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/kern.log.gz 2026-03-21T07:02:53.370 INFO:teuthology.orchestra.run.vm07.stderr:gzip -5 --verbose -- /home/ubuntu/cephtest/archive/syslog/journalctl.log 2026-03-21T07:02:53.371 INFO:teuthology.orchestra.run.vm07.stderr:/home/ubuntu/cephtest/archive/syslog/misc.log: 0.0% -- replaced with /home/ubuntu/cephtest/archive/syslog/misc.log.gz 2026-03-21T07:02:53.378 INFO:teuthology.orchestra.run.vm04.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 91.1% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-21T07:02:53.380 INFO:teuthology.orchestra.run.vm02.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 91.1% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-21T07:02:53.382 INFO:teuthology.orchestra.run.vm07.stderr:/home/ubuntu/cephtest/archive/syslog/journalctl.log: 91.2% -- replaced with /home/ubuntu/cephtest/archive/syslog/journalctl.log.gz 2026-03-21T07:02:53.383 DEBUG:teuthology.run_tasks:Unwinding manager internal.sudo 2026-03-21T07:02:53.386 INFO:teuthology.task.internal:Restoring /etc/sudoers... 2026-03-21T07:02:53.386 DEBUG:teuthology.orchestra.run.vm02:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-21T07:02:53.431 DEBUG:teuthology.orchestra.run.vm04:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-21T07:02:53.440 DEBUG:teuthology.orchestra.run.vm07:> sudo mv -f /etc/sudoers.orig.teuthology /etc/sudoers 2026-03-21T07:02:53.447 DEBUG:teuthology.run_tasks:Unwinding manager internal.coredump 2026-03-21T07:02:53.450 DEBUG:teuthology.orchestra.run.vm02:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-21T07:02:53.476 DEBUG:teuthology.orchestra.run.vm04:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-21T07:02:53.483 INFO:teuthology.orchestra.run.vm02.stdout:kernel.core_pattern = core 2026-03-21T07:02:53.484 DEBUG:teuthology.orchestra.run.vm07:> sudo sysctl -w kernel.core_pattern=core && sudo bash -c 'for f in `find /home/ubuntu/cephtest/archive/coredump -type f`; do file $f | grep -q systemd-sysusers && rm $f || true ; done' && rmdir --ignore-fail-on-non-empty -- /home/ubuntu/cephtest/archive/coredump 2026-03-21T07:02:53.490 INFO:teuthology.orchestra.run.vm04.stdout:kernel.core_pattern = core 2026-03-21T07:02:53.499 INFO:teuthology.orchestra.run.vm07.stdout:kernel.core_pattern = core 2026-03-21T07:02:53.505 DEBUG:teuthology.orchestra.run.vm02:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-21T07:02:53.538 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-21T07:02:53.538 DEBUG:teuthology.orchestra.run.vm04:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-21T07:02:53.545 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-21T07:02:53.546 DEBUG:teuthology.orchestra.run.vm07:> test -e /home/ubuntu/cephtest/archive/coredump 2026-03-21T07:02:53.549 DEBUG:teuthology.orchestra.run:got remote process result: 1 2026-03-21T07:02:53.549 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive 2026-03-21T07:02:53.552 INFO:teuthology.task.internal:Transferring archived files... 2026-03-21T07:02:53.552 DEBUG:teuthology.misc:Transferring archived files from vm02:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458/remote/vm02 2026-03-21T07:02:53.552 DEBUG:teuthology.orchestra.run.vm02:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-21T07:02:53.592 DEBUG:teuthology.misc:Transferring archived files from vm04:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458/remote/vm04 2026-03-21T07:02:53.593 DEBUG:teuthology.orchestra.run.vm04:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-21T07:02:53.601 DEBUG:teuthology.misc:Transferring archived files from vm07:/home/ubuntu/cephtest/archive to /archive/kyr-2026-03-20_22:04:26-rbd-tentacle-none-default-vps/3458/remote/vm07 2026-03-21T07:02:53.602 DEBUG:teuthology.orchestra.run.vm07:> sudo tar c -f - -C /home/ubuntu/cephtest/archive -- . 2026-03-21T07:02:53.608 INFO:teuthology.task.internal:Removing archive directory... 2026-03-21T07:02:53.608 DEBUG:teuthology.orchestra.run.vm02:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-21T07:02:53.639 DEBUG:teuthology.orchestra.run.vm04:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-21T07:02:53.644 DEBUG:teuthology.orchestra.run.vm07:> rm -rf -- /home/ubuntu/cephtest/archive 2026-03-21T07:02:53.654 DEBUG:teuthology.run_tasks:Unwinding manager internal.archive_upload 2026-03-21T07:02:53.657 INFO:teuthology.task.internal:Not uploading archives. 2026-03-21T07:02:53.657 DEBUG:teuthology.run_tasks:Unwinding manager internal.base 2026-03-21T07:02:53.659 INFO:teuthology.task.internal:Tidying up after the test... 2026-03-21T07:02:53.659 DEBUG:teuthology.orchestra.run.vm02:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-21T07:02:53.688 DEBUG:teuthology.orchestra.run.vm04:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-21T07:02:53.689 DEBUG:teuthology.orchestra.run.vm07:> find /home/ubuntu/cephtest -ls ; rmdir -- /home/ubuntu/cephtest 2026-03-21T07:02:53.691 INFO:teuthology.orchestra.run.vm02.stdout: 258075 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 21 07:02 /home/ubuntu/cephtest 2026-03-21T07:02:53.692 INFO:teuthology.orchestra.run.vm04.stdout: 258067 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 21 07:02 /home/ubuntu/cephtest 2026-03-21T07:02:53.698 INFO:teuthology.orchestra.run.vm07.stdout: 258075 4 drwxr-xr-x 2 ubuntu ubuntu 4096 Mar 21 07:02 /home/ubuntu/cephtest 2026-03-21T07:02:53.699 DEBUG:teuthology.run_tasks:Unwinding manager console_log 2026-03-21T07:02:53.704 INFO:teuthology.run:Summary data: description: rbd/iscsi/{base/install cluster/{fixed-3} conf/{disable-pool-app} supported-container-hosts$/{ubuntu_22.04} workloads/cephadm_iscsi} duration: 1128.932639837265 failure_reason: '"grep: /var/log/ceph/b16ecafc-24f1-11f1-8ede-8330751617ee/ceph.log: No such file or directory" in cluster log' flavor: default owner: kyr success: false 2026-03-21T07:02:53.704 DEBUG:teuthology.report:Pushing job info to http://localhost:8080 2026-03-21T07:02:53.728 INFO:teuthology.run:FAIL